Background
A mid-sized U.S. insurance company had been piloting GPT tools to enhance customer service and streamline claims processing. Early trials of a GPT-based virtual assistant demonstrated promise in providing quick customer answers and efficient claims review.
The insurer moved to formalize a long-term OpenAI (Azure) contract to scale these capabilities enterprise-wide. Yet, as an insurer handling sensitive personal data, they were extremely cautious. Any AI agreement would need to strictly protect client information and maintain predictable costs to satisfy regulators and the firm's risk standards.
Challenges
Reviewing the standard OpenAI service contract, the insurer uncovered several red flags:
Data Retention & Reuse Risk
The default data usage clause would have permitted the AI provider to retain and reuse the insurer's data for model training—an unacceptable risk under privacy laws and company policy.
No Spending Cap
The contract lacked any spending cap. If AI usage surged during a natural disaster or catastrophe event, costs could skyrocket with no ceiling in place.
One-Sided Liability
Liability terms were one-sided: if the AI made a bad recommendation (e.g., incorrect coverage advice), the insurer would bear all the risk with no vendor remediation support.
Lacking AI contract experience, leadership worried about signing away data control or inviting compliance and budget disasters. They decided to hire an expert to conduct an AI contract risk review before moving forward.
How Redress Compliance Helped
Redress Compliance provided a comprehensive OpenAI Contract Risk Review for the insurer. The team scrutinized the draft agreement and immediately highlighted critical issues, then negotiated targeted fixes across four key areas:
Prohibited Data Retention
Revised the contract to prohibit OpenAI from storing any customer data beyond providing the service. All policyholder data must be deleted after processing and never used to train vendor models.
Hard Monthly Spend Cap
Implemented a clause capping monthly usage charges with detailed usage reports—ensuring the firm can never be billed above an agreed-upon limit without approval. Eliminated billing surprises during peak claim periods.
Shared Responsibility
Negotiated shared responsibility if the AI's output caused a major error. If the GPT assistant gave incorrect coverage advice, the vendor would help remediate the issue instead of leaving the insurer solely liable.
Explainability Requirement
Included a requirement for the AI to provide confidence scores and rationale for its answers, giving the company oversight of automated decisions—a critical safeguard in a regulated industry.
Each change was supported by references to insurance regulations and standard industry practices, helping bring the vendor on board with the adjustments.
Need Expert AI Licensing Guidance?
Redress Compliance provides independent ai licensing advisory services — fixed-fee, no vendor affiliations. Our GenAI advisory specialists help enterprises negotiate OpenAI contracts, benchmark pricing against market rates, protect intellectual property, and build governance frameworks — typically saving 25–40% on AI spend.
Explore AI Licensing Advisory Services →Outcome and Impact
With Redress Compliance's guidance, the insurance firm secured an AI contract it could trust. What was once a risky contract became a secure, compliant partnership.
Client Testimonial
Our priority was protecting our customers and our budget—Redress Compliance made sure the AI contract did both. They caught the clauses that would have put us at risk and rewrote them in our favor. We now have an AI agreement with strong data safeguards and cost controls. Redress gave our legal and IT teams the expertise to negotiate confidently. We can innovate with AI now, knowing we're protected on all fronts.
Concerned About Hidden Risks in Your AI Contract?
Redress Compliance can be your safety net. Our AI contract risk advisory experts spot and fix fine print issues—from data use to cost exposure—before you sign.