Every day, over 200 million people type their thoughts, business strategies, medical symptoms, legal questions, and personal confessions into ChatGPT. Most of them have never read a single paragraph of OpenAI’s privacy policy. The ones who have are often left with more questions than answers, because OpenAI’s data practices are not a single, simple policy – they are a layered system of defaults, opt-outs, exceptions, and tier-dependent rules that differ radically depending on whether you are a free user, a Plus subscriber, an API customer, or an Enterprise client.
This is the full technical breakdown.
The Two OpenAIs: Consumer vs. API
The single most important distinction in OpenAI’s data architecture is the one most users never encounter: ChatGPT (the consumer product) and the OpenAI API (the developer platform) operate under fundamentally different data regimes.
ChatGPT (consumer): By default, your conversations are stored on OpenAI’s servers and can be used to train future models. This has been the default since launch, and while opt-out mechanisms exist, the default posture is data collection.
OpenAI API: Since March 1, 2023, API inputs and outputs are not used for model training by default. This policy shift – announced after significant backlash from enterprise customers – created a two-tier privacy system that persists today.
This bifurcation matters enormously. A startup founder prototyping an app with the API has fundamentally different privacy exposure than the same founder brainstorming in ChatGPT. The product looks similar. The data treatment is not.
What ChatGPT Actually Stores
When you send a message through ChatGPT’s consumer interface, OpenAI retains:
Conversation Content
Every prompt and every response is stored server-side, associated with your account. This data is retained indefinitely unless you manually delete the conversation or disable chat history. Even deleted conversations may persist in backups for up to 30 days, per OpenAI’s data retention schedule.
Account Metadata
OpenAI stores your email address, payment information (for Plus/Pro subscribers), IP address, browser type, device identifiers, and interaction timestamps. This metadata persists independently of conversation content and is governed by a separate retention policy.
Usage Telemetry
Click patterns, feature usage, session duration, model selection behavior, and error rates are all collected. This telemetry feeds product development and is explicitly carved out from any opt-out mechanism related to training data.
The March 2023 Inflection Point
On March 1, 2023, OpenAI updated its terms of service in a move that split its data practices into two distinct regimes. Before this date, API data was used for model training by default. After this date, API data was excluded from training unless customers explicitly opted in.
The consumer product did not receive the same treatment. ChatGPT free and Plus users remained in the training pipeline by default. OpenAI introduced a chat history toggle in April 2023 that allowed users to disable history storage, but the mechanism came with a critical caveat: even with history disabled, conversations are retained for 30 days for “abuse monitoring” before deletion.
This 30-day retention window is significant. It means that even users who have explicitly opted out of training still have their prompts stored, in plaintext, on OpenAI’s servers for a full month. During this window, OpenAI staff can access and review conversations flagged by automated safety systems.
The Opt-Out Mechanisms (And Their Limitations)
OpenAI offers three primary mechanisms for limiting data use:
1. Chat History Toggle
Available in ChatGPT settings, this disables persistent conversation storage and excludes your data from training. However:
- Conversations are still retained for 30 days for safety review
- OpenAI staff can access flagged conversations during this window
- Metadata collection (IP, device, timestamps) continues regardless
- The toggle must be set per-device and does not sync across sessions automatically in all configurations
2. Data Export and Deletion Requests
Users can request a full export of their data or submit deletion requests under GDPR, CCPA, or other applicable privacy regulations. OpenAI’s stated response time is 30 days, though the company reports typical fulfillment within 15 days.
3. API Data Use Opt-In
For API customers, training data use is opt-in, not opt-out. This is the inverse of the consumer default. API customers must explicitly agree to contribute data to training. The practical effect: most API-built applications never contribute to OpenAI’s training pipeline.
None of these mechanisms address metadata. OpenAI’s privacy policy explicitly reserves the right to retain and process metadata – including IP addresses, device fingerprints, and usage patterns – regardless of content opt-out status. For users concerned with zero-persistence architecture, this is the gap that policy alone cannot close.
Enterprise and Team Tier Protections
OpenAI’s Enterprise tier, launched in August 2023, introduced the strongest data protections in the company’s product line:
- No training on business data: Enterprise conversations and API usage are contractually excluded from model training. This is not a toggle – it is a contractual guarantee backed by the Enterprise terms of service.
- SOC 2 Type II compliance: OpenAI achieved SOC 2 compliance for the Enterprise product, meaning independent auditors verify controls around data handling, access, and retention.
- Data encryption at rest and in transit: AES-256 encryption for stored data, TLS 1.2+ for data in transit.
- Custom data retention windows: Enterprise customers can negotiate retention periods, including shorter windows than the default 30 days.
- SSO and admin controls: Centralized access management, audit logs, and domain verification.
The Team tier (launched January 2024, priced at $25/user/month at the time of launch) offers a subset of these protections: no training on workspace data and admin controls, but without the custom retention windows or dedicated compliance support of Enterprise.
The hierarchy is clear: free users get the least protection, Plus subscribers get slightly more, Team members get contractual training exclusion, and Enterprise clients get the full suite. Privacy, in OpenAI’s model, is a premium feature. This is precisely the dynamic that Stealth Cloud was designed to dismantle – the idea that privacy is a luxury rather than a default architectural property.
What Metadata Survives Even Full Opt-Out
This is where the analysis gets uncomfortable for privacy-conscious users. Even if you:
- Disable chat history
- Use the API with no training opt-in
- Submit a GDPR deletion request for all conversation content
OpenAI still retains:
- Account-level metadata: Email, payment records, account creation date, subscription history.
- Access logs: IP addresses, timestamps of API calls or ChatGPT sessions, geographic region derived from IP.
- Device fingerprints: Browser type, OS version, screen resolution, and language settings collected during web sessions.
- Rate limiting and abuse data: Request frequency patterns, content policy violation flags, and automated safety system outputs.
- Aggregated analytics: De-identified usage statistics that, while stripped of direct identifiers, contribute to OpenAI’s understanding of user behavior at population scale.
This metadata profile is not hypothetical. OpenAI’s privacy policy, last updated in its current substantive form in late 2024, explicitly enumerates these categories. For a company processing over 1 billion API requests per day (as of early 2025), these metadata streams represent an enormous dataset – even if not a single word of conversation content is retained.
The metadata problem is one reason why PII stripping at the client level, before data ever reaches the provider, is architecturally superior to relying on provider-side opt-outs. You cannot opt out of data that was never sent.
The Training Pipeline: How Your Data Actually Reaches a Model
When ChatGPT conversation data is used for training (the default for consumer users who have not opted out), the pipeline works roughly as follows:
- Collection: Conversations are aggregated from the production database into training datasets.
- Filtering: Automated systems attempt to remove personally identifiable information, though OpenAI acknowledges this process is imperfect.
- Annotation: A subset of conversations is reviewed by human annotators (OpenAI employees and contractors) for quality labeling and RLHF (Reinforcement Learning from Human Feedback) purposes.
- Training runs: Filtered and annotated data is incorporated into fine-tuning and training runs for future model versions.
- Model release: The resulting model weights encode statistical patterns from the training data, though individual conversations are not retrievable in their original form from the weights.
The human review step is the most privacy-sensitive. OpenAI employs teams of annotators – many of them contractors based in multiple countries – who read and label real user conversations. In January 2024, reporting revealed that OpenAI’s content moderation teams in Kenya were paid approximately $2 per hour to review graphic and disturbing content. The same pipeline that handles safety review also touches conversations flagged for quality improvement.
This means that if you are a default ChatGPT user, a human being may read your conversation. Not a statistical process. Not an abstraction. A person.
OpenAI’s Regulatory Exposure
OpenAI faces active regulatory scrutiny across multiple jurisdictions:
- Italy’s Garante: Temporarily banned ChatGPT in March 2023 over GDPR concerns, specifically around the lack of age verification and insufficient legal basis for processing personal data for training. OpenAI returned to the Italian market after implementing an age gate and more prominent opt-out disclosures.
- EU Data Protection Authorities: Multiple DPAs (including France’s CNIL and Germany’s state-level authorities) have opened investigations into ChatGPT’s data practices. The cross-border nature of these investigations, coordinated through the EDPB, could result in GDPR fines up to 4% of global revenue.
- FTC inquiries: The U.S. Federal Trade Commission opened an investigation in 2023 into whether OpenAI’s data practices constitute unfair or deceptive trade practices, with specific focus on the gap between consumer expectations and actual data handling.
- The $3 billion question: A class-action lawsuit filed in June 2023 alleged that OpenAI scraped personal data from across the internet without consent to train its models, seeking $3 billion in damages. The case remains in litigation as of early 2026.
These regulatory actions underscore a structural problem: OpenAI’s business model depends on large-scale data processing, while the global regulatory environment is moving toward data minimization. The tension is not resolvable through policy tweaks – it is architectural.
API-Specific Technical Details
For developers building on the OpenAI API, several technical details matter:
- Data retention: API inputs and outputs are retained for 30 days for abuse monitoring, then deleted. This applies to all API tiers.
- Zero data retention (ZDR): Select API customers with specific compliance needs (healthcare, finance) can negotiate zero-day retention, where API payloads are processed in memory and never written to persistent storage. This option is not publicly advertised and requires direct sales engagement.
- Fine-tuning data: If you upload data for fine-tuning, that data is stored on OpenAI’s servers for the lifetime of the fine-tuned model. Deleting the model triggers deletion of the training data, but the timeline for complete purging from backups extends to 30 days post-deletion.
- Embeddings and retrieval: Data processed through the Embeddings API follows the same 30-day retention policy as chat completions.
- Function calling and tool use: When using function calling, the function definitions, arguments, and return values are all captured within the API request and subject to the same retention policy.
The Structural Problem
OpenAI has made genuine improvements since 2022. The March 2023 API policy change, the Enterprise tier, the chat history toggle, the GDPR-responsive deletion process – these are real concessions to privacy concerns.
But they are concessions, not guarantees. Every one of these protections is revocable. They exist as policy decisions, not as architectural constraints. OpenAI can change its terms of service. It can extend retention windows. It can alter the scope of its training pipeline. The protections are legal, not mathematical.
This is the fundamental difference between policy-based privacy and architecture-based privacy. A privacy policy is a promise. Encryption is physics. A zero-knowledge proof does not require you to trust the entity verifying it. A privacy policy requires exactly that.
For organizations handling sensitive data – legal counsel, medical providers, financial advisors, journalists protecting sources – the question is not whether OpenAI’s current policies are adequate. The question is whether policy-based protection is the right category of solution at all.
The alternative is to use OpenAI’s models (which are genuinely excellent) through an architecture that strips identifying information before it reaches OpenAI’s servers, encrypts the payload client-side, processes it through ephemeral infrastructure, and ensures that even a fully compromised relay yields nothing. This is the Stealth Cloud architecture, and it works with OpenAI’s API, not against it.
Comparison: What You Get at Each Tier
| Feature | Free | Plus ($20/mo) | Team ($25/user/mo) | Enterprise (Custom) |
|---|---|---|---|---|
| Training data opt-out | Toggle | Toggle | Default off | Contractual |
| Content retention | Indefinite | Indefinite | Indefinite | Negotiable |
| Abuse monitoring retention | 30 days | 30 days | 30 days | Negotiable |
| Metadata collection | Full | Full | Full | Full (with audit) |
| Human review possible | Yes | Yes | Reduced | Minimal |
| SOC 2 compliance | No | No | No | Yes |
| Custom data processing agreements | No | No | Limited | Yes |
| SSO/Admin controls | No | No | Yes | Yes |
The Stealth Cloud Perspective
OpenAI builds excellent models and has made meaningful privacy improvements since 2022, but every protection they offer is a policy choice, not an architectural inevitability. The 30-day retention window, the metadata collection that survives every opt-out, the human review pipeline – these are features of a system where the infrastructure operator holds the keys. Stealth Cloud does not ask you to trust the operator. It removes the operator from the trust equation entirely, letting you access OpenAI’s models through zero-persistence infrastructure where the relay is architecturally blind and the only copy of your data exists in your browser.