OpenAI Negotiations

Data Privacy Risks in OpenAI Contracts

Key privacy risks in OpenAI enterprise contracts and how to mitigate them — from GDPR/CCPA compliance and data retention to fine-tuning pitfalls, indemnification, and security governance.

📅 August 3, 2025👤 Fredrik Filipsson📖 18 min read
⚠ Data training risk ⚠ Retention exposure ⚠ Fine-tuning leaks 🛡️ Mitigations available
Table of Contents

OpenAI's generative AI tools offer game-changing capabilities, but they also raise serious data privacy considerations. Global companies evaluating or negotiating OpenAI agreements must address issues from GDPR/CCPA compliance to handling sensitive business data used for model fine-tuning.

🔒

Protecting Data Privacy and Confidentiality

Enterprise contracts should explicitly safeguard all information you send to the AI and any AI-generated output. Treat this as confidential and ensure OpenAI cannot use or share your data beyond providing the service.

🚫

Secondary Use

Prohibit OpenAI from mining inputs/outputs for model training or any purpose outside your instructions.

🔐

Confidentiality

Require robust safeguards and no unauthorized disclosure to third parties.

🗑️

Retention Control

Opt for minimal retention. Secure right to request deletion on demand with certification.

Real-world warning: In 2023, Samsung discovered engineers had pasted proprietary source code into ChatGPT, effectively leaking sensitive information. This underscores the importance of strict privacy clauses and internal usage policies.
🌍

Meeting GDPR, CCPA, and Global Compliance

Global privacy regulations apply fully when you feed personal data into OpenAI. Sign a Data Processing Addendum (DPA) that spells out each party's privacy obligations — OpenAI as data processor, you as data controller.

Legal landscape: A recent U.S. court order required OpenAI to preserve all AI output logs for a lawsuit — creating tension with GDPR. Include a clause that OpenAI must inform you of any legal demands on your data and work to minimize privacy impacts.
⚗️

Model Fine-Tuning: Hidden Data Pitfalls

When you fine-tune an AI with proprietary or personal data, that information becomes part of the model's training memories. The risk: the model might regurgitate snippets of sensitive data in responses.

Best practice: Treat fine-tuning as any data-heavy project — strict data handling agreements, minimal necessary data usage, and thorough verification of results.
⚖️

Indemnification and Liability Clauses

Most OpenAI contracts offer some standard protections, but scrutinize these clauses to avoid bearing all risk.

Indemnification: Ensure OpenAI's IP indemnity is in your contract and as broad as possible. Each party should indemnify the other for risks under their control — OpenAI for AI technology issues, you for misuse of the service.

Liability limits: OpenAI likely caps liability and excludes indirect damages. A strict cap may be unacceptable for mission-critical tasks. Negotiate this. Aim for carve-outs from the cap:

Critical: Verify that indemnification obligations from OpenAI are outside the liability cap. Consider requiring OpenAI to carry cyber liability insurance. The goal: avoid bearing all financial pain for mistakes outside your control.
🛡️

Data Security and Governance

Risk Mitigation Summary

Risk AreaMitigation in Contract & Practice
Unauthorized data useExplicit "no-training" clause. Data solely for your organization's service.
GDPR/CCPA non-complianceSign DPA. Include deletion obligations, data subject request assistance, EU SCCs.
Over-retention of dataNegotiate retention limits. Leverage Enterprise features for minimal retention.
Data breach liabilityCarve out breaches from liability limits. Require prompt notification and cooperation.
Fine-tuning leaksEnsure exclusive use. Test models for leakage. Limit PII in training datasets.
📋

Recommendations

1

Demand a robust DPA

Execute OpenAI's Data Processing Addendum covering GDPR, CCPA, breach notification, and compliance request assistance.

2

Lock down data use in contract

Add clear confidentiality clauses. Specify OpenAI may only use your data to provide service — no sharing, no secondary use.

3

Set data controls and retention

Choose zero-retention for sensitive inputs. Ensure you can request data deletion at any time with swift compliance.

4

Carve out critical liabilities

Push to carve data breaches, confidentiality breaches, and IP indemnity from liability caps. Get higher or uncapped protection.

5

Secure IP indemnity and more

Ensure indemnification for IP claims. Discuss coverage for privacy violations or defamation. Even partial agreement highlights your concerns.

6

Insist on security assurances

Confirm industry security standards, incident notification obligations, and assistance with security investigations.

7

Prepare internal guidelines

Develop rules for employees — prohibit entering customer PII or secret source code. Control what goes into the model to reduce breach risk.

8

Evaluate the need to share

Before sending any dataset (especially for fine-tuning), evaluate if it's truly necessary. Share minimum data required. Less data = lower risk.

Checklist: 5 Actions to Take

1Audit Your Data: Inventory data types you plan to send. Classify what's sensitive. Remove or anonymize high-risk data before it ever reaches the AI.
2Get Paperwork in Place: Request and sign OpenAI's DPA for GDPR/CCPA. Sign additional agreements if needed (e.g., HIPAA BAA). Attach to your main contract.
3Negotiate Key Terms: Review the standard contract for gaps. Propose amendments on data use, confidentiality, retention/deletion, indemnities, and liability carve-outs.
4Establish Usage Policies: Create clear internal policy on what information is off-limits. Train employees about privacy and security risks.
5Monitor and Adapt: Continuously monitor compliance. Ensure OpenAI honors deletion requests. Monitor outputs for data exposure. Update practices as regulations evolve.

FAQ

Q1

Can we use OpenAI without violating GDPR or CCPA?

Yes, with compliance steps. Have OpenAI sign a DPA committing to GDPR/CCPA principles. Utilize data retention controls. Avoid inputting personal data unless necessary with proper legal basis. With proper contract and configuration, OpenAI can be used in line with global privacy laws.

Q2

Will OpenAI use our data to train its models?

For enterprise and API users, OpenAI does not use your data to train general models by default. Your prompts and outputs stay isolated. Confirm this explicitly in your contract — ensure it states OpenAI won't use your data for research or improvement without permission.

Q3

What happens if OpenAI has a data breach involving our information?

OpenAI should inform you immediately. You'll manage impact on your side (notifying individuals/regulators). Negotiate liability so that if the breach was OpenAI's fault, they cover costs like regulatory fines and customer notifications. Don't accept overly strict liability limits for breaches.

Q4

How do we handle highly sensitive or regulated data?

Cautiously and with extra safeguards. Check if special agreements are needed (HIPAA BAA for health data). Consider specialized cloud regions (Azure OpenAI) for jurisdictional requirements. Minimize what you share — every piece of sensitive data that stays out of the system is one less piece that could be exposed.

Q5

Do we own the AI's outputs and our data?

Yes — you retain ownership of both inputs and outputs. Ensure the contract explicitly states all inputs and outputs are your confidential information with full rights. This lets you use results freely in your business. OpenAI must treat outputs with the same care as any sensitive data.

Read about our GenAI Negotiation Service.

Read about our OpenAI Contract Negotiation Case Studies.

Need Help Securing Your OpenAI Contract?

Redress Compliance provides independent GenAI advisory — from data privacy risk assessment and contract redlining to compliance strategy and negotiation support.