Gartner predicted that by 2024, 60% of the data used for AI and analytics projects would be synthetically generated. The actual figure was closer to 35%, but the trajectory was unmistakable. The synthetic data market reached $1.1 billion in 2025, driven by a compelling promise: you can train AI models on data that looks real, behaves statistically like real data, but was never collected from any actual person. Privacy problem solved.
Except it isn’t.
Synthetic data occupies a peculiar position in the AI privacy landscape. It is simultaneously one of the most promising technical approaches to privacy-preserving AI development and one of the most overhyped. The gap between what synthetic data can do in theory and what it delivers in practice is substantial, and the consequences of that gap fall on the people whose privacy synthetic data is supposed to protect.
Understanding why requires examining the technical mechanics of synthetic data generation, the specific failure modes that undermine its privacy guarantees, and the organizational incentives that drive its adoption even when those guarantees are weak.
How Synthetic Data Works
Synthetic data is generated by AI models trained on real data. The generative model learns the statistical properties, distributions, correlations, and patterns present in the original dataset, then produces new data points that preserve those statistical properties without copying any individual’s actual records.
Generation Methods
The primary approaches to synthetic data generation include:
Generative Adversarial Networks (GANs): Two neural networks – a generator and a discriminator – compete in a training loop. The generator creates synthetic data, the discriminator tries to distinguish synthetic from real. Through iterative training, the generator learns to produce data that is statistically indistinguishable from the original. GANs are particularly effective for generating synthetic images, tabular data, and time-series data.
Variational Autoencoders (VAEs): These models learn a compressed latent representation of the real data distribution, then sample from this latent space to generate new data points. VAEs provide more control over the generation process and are commonly used for structured data synthesis.
Large Language Models (LLMs): GPT-class models can generate synthetic text data, survey responses, clinical notes, and other text-based datasets. The LLM is prompted or fine-tuned to produce data that mimics the characteristics of a target dataset.
Statistical methods: Copula-based approaches, Bayesian networks, and other statistical frameworks model the joint probability distribution of the real data and sample from it to produce synthetic records. These methods are less flexible than neural approaches but more interpretable and easier to audit.
Differential privacy synthetic data: A mathematically rigorous approach that injects calibrated noise during the generation process to provide formal privacy guarantees. Mechanisms like PATE-GAN and DP-SGD augmented generators provide provable bounds on the maximum information any synthetic record can reveal about any individual in the training data.
The Quality-Privacy Trade-Off
Every synthetic data generation method embodies a fundamental trade-off between data utility and privacy protection. The more faithfully the synthetic data replicates the statistical properties of the real data, the more useful it is for downstream tasks – but also the more vulnerable it is to privacy attacks that recover information about the real data.
This trade-off is not merely practical; it is mathematical. Research by Stadler, Oprisanu, and Troncoso (2022) formally proved that synthetic data generators face an inherent tension: achieving high utility requires preserving fine-grained patterns in the data, but those fine-grained patterns may encode information about specific individuals in the training set.
The trade-off sharpens for rare or outlier records. A synthetic data generator that accurately reproduces the statistical properties of a dataset containing rare medical conditions must, by definition, capture patterns that are associated with a small number of real patients. The rarer the condition, the more the synthetic pattern points back to real individuals.
The Privacy Failures of Synthetic Data
The marketing narrative around synthetic data positions it as inherently private – after all, the synthetic records were never collected from anyone. The research literature tells a different story.
Membership Inference Attacks
Membership inference attacks attempt to determine whether a specific individual’s data was in the training set used to generate the synthetic data. If successful, the attack reveals that the individual contributed data to the original dataset – a privacy violation in itself, and a foundation for further attacks.
A 2023 study by researchers at University College London tested five leading commercial synthetic data generators against membership inference attacks, the results of which contributed to the growing body of evidence that privacy compliance in AI remains fundamentally challenging. The attack success rate ranged from 62% to 89% depending on the generator and dataset – far above the 50% baseline of random guessing. For datasets with rare or outlier records, success rates exceeded 95%.
The implication is direct: synthetic data generated from a medical dataset, for example, can reveal with high confidence whether a specific patient was in the original dataset. This is a meaningful privacy violation, particularly for sensitive datasets where membership itself is informative (e.g., a dataset of patients with a specific diagnosis, or a dataset of individuals under criminal investigation).
Attribute Inference Attacks
Attribute inference attacks go further: given partial information about an individual known to be in the training data, the attacker uses the synthetic data to infer unknown attributes. If the synthetic data faithfully preserves the correlations between attributes in the real data, knowing some of an individual’s attributes can reveal the rest through statistical inference.
Research published at the 2024 IEEE Symposium on Security and Privacy demonstrated attribute inference attacks against synthetic tabular data that recovered sensitive attributes (income, health status, credit score) with 73-91% accuracy for individuals whose partial records were known to the attacker. The accuracy increased for datasets with strong inter-attribute correlations – precisely the datasets where synthetic data is most useful because it preserves those correlations for downstream analysis.
Reconstruction and Memorization
Synthetic data generators can memorize and reproduce specific training records, particularly when the generative model is large relative to the training dataset or when certain records are statistically unusual.
A 2024 study in Nature Machine Intelligence examined synthetic health data generated by multiple commercial platforms and found that 3.7% of synthetic records were near-exact replicas of real patient records in the training data. For patients with rare conditions or unusual combinations of attributes, the replication rate rose to 12.4%. These near-replicas were not “synthetic” in any meaningful privacy sense – they were copies of real people’s health data with trivial perturbations.
This memorization problem parallels the model memorization issues in large language models, where training data can be extracted from model outputs. The generative model used to create synthetic data is itself a model that has memorized aspects of its training data, and its outputs (the synthetic dataset) can leak that memorized information.
The Differential Privacy Question
Differential privacy (DP) is the gold standard for mathematical privacy guarantees. Applied to synthetic data generation, DP ensures that the inclusion or exclusion of any single individual’s record in the training data changes the probability of any synthetic output by at most a small, quantifiable amount (controlled by the privacy parameter epsilon).
The Promise
DP-synthetic data provides something no other approach offers: a provable, mathematical bound on privacy risk. If the epsilon parameter is sufficiently small, no attack – however sophisticated – can extract meaningful information about any individual from the synthetic data. This guarantee is universal and information-theoretic; it does not depend on assumptions about the attacker’s capabilities or knowledge.
The Practice
In practice, DP-synthetic data faces severe utility challenges. The noise injection required to achieve strong privacy guarantees (low epsilon values) degrades data quality substantially. A 2024 benchmarking study by the National Institute of Standards and Technology (NIST) found that:
- At epsilon = 1 (strong privacy), synthetic tabular data preserved only 40-60% of the statistical relationships present in the original data
- At epsilon = 10 (moderate privacy), preservation improved to 70-85% but privacy guarantees weakened to a level that many privacy researchers consider insufficient
- Achieving both high utility and strong privacy required datasets with at least 100,000 records; for smaller datasets, the noise dominated the signal
The consequence is that DP-synthetic data works well for large datasets with broad statistical patterns, but poorly for the detailed, fine-grained, or small-population datasets that are often the most privacy-sensitive.
The Epsilon Problem
The privacy guarantee of differential privacy is only as strong as the epsilon parameter chosen. There is no universal standard for what constitutes a “safe” epsilon value. In practice, organizations face pressure to increase epsilon (weaken privacy) to maintain data utility, and the resulting epsilon values often provide privacy guarantees that are mathematically well-defined but practically weak.
A survey of DP-synthetic data deployments published at the 2024 ACM Conference on Computer and Communications Security found that 67% used epsilon values above 10, and 23% used values above 100. At epsilon = 100, the differential privacy guarantee is effectively meaningless – the noise is too small to prevent privacy attacks against outlier records.
The organizational incentive structure is clear: teams generating synthetic data are evaluated on utility metrics (how well the synthetic data serves downstream tasks) rather than privacy metrics (how resistant the synthetic data is to privacy attacks). When utility and privacy trade off, utility wins.
The Regulatory Blind Spot
Synthetic data occupies an ambiguous position in data protection regulation, and that ambiguity is being exploited.
GDPR and Synthetic Data
The GDPR applies to “personal data” – information relating to an identified or identifiable person. If synthetic data truly contains no information about real individuals, it falls outside GDPR’s scope. This is the regulatory promise that drives much of the commercial interest in synthetic data: generate synthetic data from personal data, then use the synthetic data freely without GDPR constraints.
The European Data Protection Board (EDPB) has not issued definitive guidance on when synthetic data qualifies as anonymous (and thus outside GDPR scope). The Article 29 Working Party’s Opinion on Anonymization Techniques (2014) established that anonymization must be irreversible and must resist reidentification attacks including linkage, inference, and singling out. Whether synthetic data meets these criteria depends on the generation method, the privacy parameters, and the characteristics of the underlying real data.
The risk for organizations is significant: if a regulator or court determines that a specific synthetic dataset does not qualify as anonymous under GDPR – because it is vulnerable to membership inference, attribute inference, or reconstruction attacks – then the entire downstream use of that dataset has occurred without a lawful basis for processing. The GDPR compliance implications cascade through every system that consumed the synthetic data.
The “Laundering” Concern
Privacy researchers have raised concerns that synthetic data can function as a data laundering mechanism: personal data that cannot legally be used for AI training is converted into synthetic data that is treated as non-personal, and the synthetic data is then used for the originally prohibited purpose.
The laundering concern is particularly acute for data collected without adequate consent. If an organization collected user data under a privacy policy that did not authorize AI training, generating synthetic data from that collection and using the synthetic data for training achieves the same outcome that the original policy prohibited – but through a technical intermediary that claims to sever the connection to real individuals.
Whether this severance is genuine depends on the technical properties of the synthetic data. As the privacy attack research demonstrates, the severance is often weaker than the generating organization claims.
When Synthetic Data Works
Despite its limitations, synthetic data provides genuine privacy value in specific contexts:
Software testing and development. Synthetic data for testing database schemas, software interfaces, and processing pipelines provides realistic data structures without exposing real user information. The privacy risk in this context is low because the synthetic data doesn’t need to preserve the fine-grained statistical properties that enable privacy attacks – it just needs to look structurally realistic.
Data augmentation for underrepresented groups. Generating synthetic examples of rare conditions, minority demographics, or unusual patterns can improve model performance without oversampling real individuals. This use case improves both fairness and privacy when implemented carefully.
Tabletop exercises and demonstrations. Synthetic data for security training, compliance demonstrations, and product demos provides realistic scenarios without exposing real data to audiences who shouldn’t access it.
Federated and distributed contexts. Synthetic data can serve as a privacy-preserving communication mechanism in federated learning architectures, where participants share synthetic representations of their local data rather than the data itself.
In each of these cases, the key success factor is that the downstream use does not require the synthetic data to faithfully preserve individual-level patterns in the real data. When fidelity requirements are low, privacy protection is strong. When fidelity requirements are high, the trade-off reasserts itself.
The Honest Assessment
Synthetic data is a useful privacy engineering tool, not a privacy solution. The distinction matters.
A tool reduces risk in specific contexts when applied with appropriate expertise and realistic expectations. A solution eliminates the problem. Synthetic data does not eliminate the privacy problem in AI development – it redistributes and partially mitigates it.
The organizations marketing synthetic data as a privacy panacea have a financial interest in overstating its protections. The organizations adopting synthetic data to satisfy regulatory requirements have an incentive to accept those overstated claims without rigorous technical validation. The result is an ecosystem where synthetic data is deployed with confidence that exceeds its actual privacy guarantees, and where the residual risk falls on the individuals whose real data seeded the generation process.
A genuinely privacy-preserving approach to AI does not require trust in the statistical properties of a generated dataset. It requires an architecture where private data never persists, never aggregates, and never becomes available for any use beyond the immediate purpose for which it was provided.
The Stealth Cloud Perspective
Synthetic data attempts to solve a problem that zero-knowledge architecture eliminates. The question “how do we use private data safely?” presupposes that private data must be collected, stored, and repurposed. Stealth Cloud starts from a different premise: private data should not exist outside the moment of its use. When conversations are processed in volatile memory and cryptographically shredded upon completion, there is no dataset from which to generate synthetic derivatives, no training pipeline to feed, and no privacy trade-off to optimize. Synthetic data is a patch on a leaking architecture. Zero persistence is a pipe that never leaks.