Note on method and confidentiality: This review synthesises and organises the range of arguments, concerns, and proposals debated during the symposium. The event operated under Chatham House rules. Accordingly, no participants are named, no organisations are identified, and no remarks are attributed. Where it is helpful to convey the texture of the discussion, the review uses formulations such as “some participants argued” or “others cautioned” to reflect the existence of contrasting views without identifying sources.
1. Purpose and framing
The symposium convened practitioners, methodologists, privacy specialists, governance leaders, and social-science researchers to interrogate the ethical landscape of synthetic market research. The central focus was the use of (i) synthetic personas as population- or segment-level representations and (ii) digital twins as more granular, sometimes longitudinal, behavioural simulations. Participants broadly agreed that synthetic research is rapidly shifting from a novelty used for ideation toward an operational capability used to shape product strategy, pricing, messaging, and even organisational policy. This shift elevates the ethical stakes because synthetic systems increasingly influence decisions that affect real people at scale.
From the outset, the symposium adopted a practical working definition: synthetic market research refers to methods that generate insights by simulating human responses or behaviours using computational models calibrated on empirical data, rather than directly sampling and surveying humans for each study. The symposium then treated “ethics” not as an afterthought but as a methodological property: the quality of an insight depends on its epistemic legitimacy, its social acceptability, and the integrity of the processes that produced it.
Several participants argued that the ethical debate is often distorted by an overly binary framing: either synthetic research is “fake” and therefore unacceptable, or it is “privacy-preserving” and therefore inherently ethical. The symposium rejected both extremes. A recurring theme was that synthetic methods sit inside a set of trade-offs that must be managed explicitly: utility versus privacy, speed versus validity, and commercial advantage versus social harm. The most productive conversations treated these trade-offs as design variables that can be measured, bounded, disclosed, and governed.
2. Taxonomy: what exactly is being simulated?
A significant portion of the symposium focused on disentangling categories that are frequently conflated in marketing discourse: synthetic data, synthetic respondents, synthetic personas, and digital twins. The discussion stressed that ethical risk is not uniform across these categories. Ethics is shaped by the object of simulation, the granularity of representation, and the use context.
- Synthetic data was characterised as generated datasets intended to preserve statistical properties of an original dataset (for example, transaction logs, demographic tables, or survey microdata) while reducing disclosure risk. Participants emphasised that synthetic data has a long history in statistics and privacy research, but new generative methods can make it both more useful and more dangerous if misapplied.
- Synthetic respondents were described as interactive systems (often LLM-based) that answer questions in a survey-like format. Discussion highlighted that these systems can produce fluent, persuasive outputs that may conceal uncertainty and fabricate plausible details.
- Synthetic personas were treated as structured profiles representing archetypes or segments. Participants noted that personas can function as “interfaces” to complex distributions, but they can also embed stereotypes if built uncritically.
- Digital twins were characterised as higher-resolution models sometimes intended to simulate an individual or micro-cohort over time. The symposium repeatedly returned to the claim that “twinhood” is not a binary property. A model may be more or less twin-like depending on its calibration data, its behavioural fidelity, and the specificity of its intended use.
A key consensus emerged: as systems move from population-level simulation toward person-like simulation, ethical complexity increases nonlinearly. Even participants optimistic about synthetic research acknowledged that high-resolution twins can create a moral “shadow” problem: the system becomes a proxy that can be queried, interrogated, and optimised against in ways the underlying person never consented to and may not benefit from.
3. Foundational ethical principles and the “research integrity” lens
While the symposium drew on familiar responsible AI frameworks (human-centric values, transparency, fairness, robustness, accountability), the strongest thread was a research-integrity lens grounded in professional market research norms: clarity of method, honest reporting, avoidance of deception, and respect for subjects. Participants argued that synthetic research requires an additional discipline: epistemic humility. Because synthetic systems can generate confident-looking answers at low marginal cost, there is a structural incentive to ask more questions than the system can answer reliably, and then to treat its outputs as empirical observations.
Some participants proposed a simple ethical test: “Would a competent reviewer, given the method description, accept the claims as justified?” This reframed ethics as a form of falsifiability and auditability. Under this view, the primary ethical failure is not merely privacy leakage or bias, but the production of actionable claims without adequate grounding, especially when those claims may harm vulnerable groups.
Others challenged the sufficiency of integrity norms alone, arguing that market research ethics historically evolved for human-subject research and may not fully address model-mediated risks: inversion attacks, membership inference, model drift, or emergent manipulative optimisation. The symposium therefore treated research ethics and AI safety as complementary rather than competing domains.
4. Privacy: synthetic is not automatically anonymous
Privacy was the most intensely debated theme. Participants aligned on a foundational point: synthetic outputs do not guarantee anonymity. The presence or absence of privacy risk depends on the generation process, the uniqueness of records, the adversary model, and the amount of auxiliary information an attacker may possess.
Several participants stressed that “synthetic” is often used rhetorically to imply “non-personal.” This is a category error. A synthetic record may still encode information that allows re-identification or membership inference if the generator overfits or if the dataset contains rare combinations. This is especially salient for high-dimensional behavioural data (location trails, purchase sequences, or free-text responses) where uniqueness is common.
In response, others argued that synthetic methods can still deliver meaningful privacy gains when engineered correctly. The discussion distinguished between:
- De-identification by transformation (masking, generalisation), which can erode utility and still fail under linkage attacks.
- Distributional synthesis, which aims to preserve aggregate properties without reproducing individual records.
- Formal privacy approaches (for example, differential privacy or other provable methods), which bound disclosure risk but require careful parameterisation and can reduce fidelity.
The symposium did not settle on a single “best” approach. Instead, participants converged on a pragmatic proposition: privacy must be tested, not presumed. Synthetic research programmes should incorporate explicit adversarial evaluation, including membership inference-style probes, uniqueness checks, and red-team exercises appropriate to the sensitivity of the source data and the stakes of deployment.
5. Consent, purpose limitation, and the ethics of calibration
A major question ran through multiple sessions: when a synthetic system is calibrated on real human data, what are the ethical obligations to those humans? Participants distinguished between legal compliance and ethical legitimacy. Even if a dataset was collected under a lawful basis (contract, legitimate interest, or consent), the ethical acceptability of using it to build simulation systems may depend on expectations and downstream use.
Some participants argued for a strict interpretation: if individuals did not meaningfully understand that their data could be used to build models that simulate their preferences, it is ethically questionable to do so, especially for twins that approximate individual-level behaviour. Others proposed a more contextual view: calibration is ethically acceptable if the system is used for population-level insights, includes strong privacy protections, and is not used to make decisions about specific individuals.
The symposium introduced the idea of “purpose distance”: the greater the distance between the original purpose of collection and the new modelling purpose, the stronger the duty to re-consent, disclose, or restrict. Several participants proposed that digital-twin uses should be treated as higher-purpose-distance by default, unless the twin is explicitly authorised as part of a user-facing product (for example, a personal assistant that a user chooses to configure).
There was vigorous discourse on whether “twin-like” simulation can be ethically justified for customer research without explicit consent. Advocates claimed it is analogous to predictive analytics and segmentation already used in marketing, merely with a richer interface. Critics countered that twin systems change the moral nature of inference: they create a persistent, queryable proxy that can be interrogated for latent traits, including sensitive inferences, at far lower cost and with far greater granularity than conventional analytics.
6. Transparency, non-deception, and disclosure obligations
Participants repeatedly returned to the ethics of representation: how synthetic insights are communicated to clients, stakeholders, and end users. The concern was not merely reputational. It was epistemic and moral. If synthetic outputs are framed as if they were derived from human respondents, they can mislead decision-makers, distort accountability, and erode trust in research as a discipline.
A widely supported recommendation was a disclosure standard that distinguishes:
- Empirical observations (measured from real humans in a defined sample).
- Model-based estimates (predictions or simulations derived from calibration data).
- Hypothesis-generation outputs (directional insights intended to guide further validation).
Some participants argued for a “nutrition label” approach: a compact, standardised method card appended to synthetic studies that specifies data sources, calibration time window, intended domain of validity, uncertainty indicators, and prohibited uses. Others cautioned that overly technical disclosure can become performative, a box-ticking exercise that does not change behaviour. The symposium therefore debated how to make disclosure both comprehensible and enforceable, particularly under time pressure in commercial settings.
An additional nuance concerned interactive systems. When synthetic respondents are deployed in workshops or product teams, they can be psychologically experienced as “people.” Multiple participants warned that this anthropomorphic pull can create unintentional deception even if the system is nominally labelled synthetic. Some proposed interface-level mitigations: persistent disclosures, uncertainty visualisation, and interaction design that discourages over-identification with outputs.
7. Validity, reliability, and the ethics of over-claiming
The symposium’s most “scientific” exchanges centred on validity. Participants treated invalid inference as an ethical harm because it can drive decisions that disadvantage consumers, employees, or communities. The discussion separated several layers of validity:
- Internal validity: do synthetic outputs correctly reflect patterns in the calibration data?
- External validity: do outputs generalise to the real world outside the training distribution?
- Construct validity: are the simulated constructs (attitudes, intentions, willingness-to-pay) meaningfully mapped to real-world behaviours?
- Predictive validity: do synthetic findings predict future outcomes when tested empirically?
Some participants emphasised that synthetic systems are best suited for “fast iteration” rather than final measurement: exploring hypotheses, stress-testing messaging, or identifying potential segmentation variables. Others argued that for certain domains with rich historical data and stable behaviours (some categories of repeat purchase or certain pricing patterns), model-based simulation can be decision-grade if validated continuously.
Critics of decision-grade claims highlighted a recurrent risk: synthetic respondents can reproduce the “commonsense” story of a market rather than the market itself. They may articulate coherent rationales that align with cultural narratives yet fail to predict behaviour. This was framed as a form of narrative overfitting. In response, proponents advocated for benchmark designs: running synthetic studies in parallel with human studies, measuring divergence systematically, and using divergence as a diagnostic signal rather than an embarrassment to be hidden.
A further debate concerned uncertainty. Many participants argued that synthetic outputs should be accompanied by uncertainty estimates analogous to confidence intervals, even if derived from model ensembles, resampling, or sensitivity analysis rather than classical sampling theory. Others cautioned that uncertainty quantification in generative systems is still immature and can be misleading if presented with false precision. The symposium leaned toward a middle ground: use uncertainty indicators as decision aids, but disclose their method and limits, and prioritise calibrated back-testing where feasible.
8. Bias, fairness, and representational harm
Bias and fairness were discussed as both statistical and moral phenomena. Participants noted that synthetic systems can encode bias from the calibration dataset, amplify it through model priors, and then launder it under a veneer of neutrality. Because synthetic outputs are often delivered as articulate prose, they can appear more authoritative than noisy human responses, increasing the risk that biased patterns become institutionalised.
Representational harm received particular attention. Even if a system does not expose personal data, it can still harm groups by stereotyping them, reducing their diversity to caricature, or ignoring minority experiences. Participants discussed scenarios where a synthetic persona for a demographic group becomes “the story” used by product teams, displacing actual engagement with members of that group.
Several participants advocated for fairness testing that mirrors clinical thinking: subgroup performance must be evaluated explicitly, and the system’s “coverage” should be mapped. This includes identifying populations where the model is unreliable due to sparse data, cultural mismatch, or dynamic social context. Others urged caution, arguing that subgroup labels can themselves be sensitive and that measurement can inadvertently reify categories. The symposium did not resolve this tension, but it did produce a shared direction: ethical synthetic research requires both quantitative evaluation and qualitative review by people with contextual expertise, particularly for marginalised groups.
9. Manipulation, persuasion, and the moral hazard of optimisation
One of the most forceful sessions focused on the downstream use of synthetic research to optimise persuasion. Participants acknowledged that marketing has always sought to influence behaviour. The concern is that synthetic systems can industrialise the process: rapid experimentation on simulated minds can produce messaging strategies designed to exploit cognitive vulnerabilities, target subpopulations, and fine-tune nudges at scale.
Some participants argued that this risk is not unique to synthetic methods; it is the same moral problem found in A/B testing and behavioural advertising. Others responded that simulation materially changes the incentive landscape: it lowers the cost of exploring manipulative strategies, increases the speed of iteration, and can generate persuasive rationales that make questionable tactics appear “validated.” Under this view, synthetic research can become an accelerant for ethically problematic influence campaigns.
The symposium explored guardrails. Proposals included:
- Use restrictions for certain categories: minors, sensitive traits, health, financial distress, addiction-related products, and other vulnerability contexts.
- Prohibited objectives: designing misinformation, suppressing consumer autonomy, or exploiting known cognitive weaknesses.
- Internal review gates for high-impact campaigns, analogous to ethics boards in human-subject research.
- Audit trails that record what questions were asked, which prompts were used, and how outputs influenced decisions.
There was debate on enforceability. Some participants doubted that “principles” without accountability would matter. Others argued that governance can be operationalised through contracts, model usage policies, and monitoring. A recurring observation was that ethical alignment must be embedded in incentives: if teams are rewarded solely for conversion, they will treat guardrails as friction unless leadership treats them as non-negotiable constraints.
10. Security, abuse resistance, and adversarial environments
Security discussions were notable for connecting technical failure modes with moral outcomes. Participants emphasised that a synthetic research stack is a compound system: data pipelines, prompt templates, model APIs, storage layers, and user interfaces. A breach or leakage in any layer can become a privacy harm. Moreover, even without a classical breach, generative systems can be attacked through prompt injection, exfiltration prompts, or data reconstruction attempts.
Some participants described these risks as “research security.” In traditional market research, confidentiality failures are reputational and contractual. In synthetic research, confidentiality failures can also be computational: a model can be induced to reveal sensitive patterns or training artefacts. In response, the symposium debated operational controls: access segmentation, key management, red-team testing, rate limits, monitoring of anomalous queries, and separation between calibration environments and production analysis.
Participants also raised the issue of insider misuse. Synthetic systems can allow employees to query simulated populations for sensitive inferences. Even if no real individual is directly identifiable, the ability to ask “what would a vulnerable person do” or “how do we exploit reluctance” can be morally corrosive. Several participants argued that responsible synthetic research must treat misuse as a predictable scenario, not an edge case, and build monitoring and governance accordingly.
11. Intellectual property, ownership, and rights over “likeness”
Ownership debates were more complex than anticipated. Participants discussed multiple layers of value: the raw calibration data, the synthetic population artefact, the model configuration, the prompts and instruments, and the derived insights. Who owns each layer is often unclear in practice, particularly when vendors combine client data with proprietary models.
Ethically, the most charged question was whether there are moral rights in a person’s “likeness” when a model is calibrated on their behaviour. Some argued that if individuals can be simulated in a way that approximates their preferences and vulnerabilities, there is a strong moral case for consent and at least some control, regardless of legal status. Others urged restraint, noting that “likeness” can be an ambiguous concept and that overextending it might impede legitimate aggregate modelling. The symposium did not settle the issue, but participants agreed it is a looming area of contention as digital-twin claims become more concrete.
Practically, participants advocated for clearer contracts and customer-facing commitments: boundaries on reuse, restrictions on resale, explicit retention periods, and clear statements about whether client data is used to improve shared models.
12. Societal impacts: trust, displacement, and epistemic inequality
Beyond immediate technical and organisational ethics, the symposium considered second-order societal effects. A prominent concern was trust. If synthetic research is marketed as “as good as” human research without disclosure, and if it generates notable failures, it may erode public trust in research more broadly. Participants argued that the insights industry relies on credibility that can be lost quickly and is difficult to rebuild.
Another theme was displacement and voice. If synthetic substitutes become dominant, fewer real people may be asked for their opinions, particularly in early-stage product development where budgets are limited. Some participants viewed this as ethically troubling: public discourse and product design could become increasingly shaped by simulations rather than lived experience. Others countered that synthetic tools can expand access by lowering costs and enabling more frequent testing, thereby increasing responsiveness to consumer needs. The debate turned on implementation: whether synthetic is used as a complement to human engagement or a replacement for it.
The symposium also raised the concept of epistemic inequality. Organisations with powerful synthetic tools may accumulate an advantage in predicting and shaping consumer behaviour, widening gaps between large firms and smaller competitors. Participants discussed whether this dynamic could drive more manipulation, reduce market diversity, or concentrate influence. While the symposium did not propose policy solutions, it did identify this as a topic warranting continued attention.
13. Areas of convergence: where participants largely agreed
Despite vigorous disagreement on many points, several areas of convergence emerged clearly:
- Ethics is not separable from method. Over-claiming and poor validation are ethical problems because they can cause real harm.
- Synthetic is not automatically private. Privacy risk must be assessed empirically with explicit threat models.
- Disclosure is essential. Synthetic outputs should not be presented as if they were directly sampled from humans.
- Granularity increases risk. More twin-like systems require stronger governance than segment-level personas.
- Guardrails must be operational. Principles alone are insufficient; programmes need policies, audit trails, and accountability.
Participants also agreed that the field needs shared vocabulary. Terms like “digital twin” are currently used inconsistently, enabling marketing claims that outrun methodological reality. A proposed remedy was a tiered classification scheme describing how a system is calibrated, what it can credibly simulate, and what uses are prohibited.
14. Persistent tensions: where disagreement remained unresolved
The symposium’s most enduring tensions were productive rather than paralysing. They can be summarised as disputes about thresholds, not absolutes:
- Consent thresholds: Is explicit consent necessary for any twin-like calibration, or can aggregate legitimate-interest approaches be ethically acceptable with strong protections?
- Decision-grade claims: Under what validation regime, and for which domains, can synthetic research substitute for human studies?
- Fairness measurement: How can subgroup evaluation be conducted without reifying sensitive categories or increasing privacy risk?
- Manipulation boundaries: Where is the line between legitimate persuasion and unethical exploitation, and who sets it in commercial contexts?
- Disclosure design: What disclosure format changes behaviour rather than becoming ritualised compliance?
Several participants argued that these tensions cannot be solved by abstract principle. They require case law in practice: shared examples, post-mortems of failures, and an iterative professional consensus similar to how other applied fields mature.
15. Emerging proposals: toward a “responsible synthetic research” programme
By the final sessions, discussion shifted from diagnosis to constructive design. Multiple proposals converged into an implicit programme architecture. While details varied, the composite shape was consistent:
15.1 Governance and accountability
- Define roles and escalation paths for ethics and risk review.
- Create an approval gate for high-impact uses (vulnerable populations, sensitive categories, or high-stakes decisions).
- Maintain an auditable record of study configuration: data sources, prompts, model versions, and reporting outputs.
15.2 Technical and methodological controls
- Adopt routine privacy testing and red-teaming aligned to explicit threat models.
- Implement validation benchmarks against human studies, including divergence tracking over time.
- Use sensitivity analysis to assess how outputs change under plausible shifts in assumptions or calibration data.
15.3 Reporting standards
- Include a standard “method card” that discloses synthetic vs empirical components, calibration windows, and limitations.
- Separate hypothesis-generation outputs from measurement claims, and label them distinctly.
- Provide uncertainty indicators with transparent explanation of their derivation and limitations.
15.4 Use policy and ethical boundaries
- Define prohibited objectives (misinformation, exploitation of vulnerability, discriminatory targeting).
- Define restricted domains requiring additional oversight.
- Ensure contractual commitments regarding data use, retention, and model improvement.
Some participants recommended that the industry should converge on an independent assurance mechanism. Ideas ranged from voluntary certification schemes to peer-review style audits. Skeptics questioned whether certification can keep pace with fast-evolving models, but proponents argued that even imperfect assurance can raise baseline practice and create market pressure for transparency.
16. Research agenda: questions the symposium prioritised
The symposium concluded with a forward-looking discussion of open research needs. The following topics were repeatedly prioritised:
- Uncertainty quantification for generative respondents: practical, calibrated uncertainty measures that are not merely rhetorical.
- Ground-truth benchmarking designs: efficient study designs that measure where synthetic methods succeed or fail across domains.
- Privacy evaluation standards: widely accepted tests for disclosure risk in synthetic datasets and synthetic respondent systems.
- Bias measurement and mitigation: methods that address both statistical bias and representational harm in narrative outputs.
- Human factors and interface ethics: how interaction design shapes user trust, anthropomorphism, and over-reliance.
- Governance effectiveness: empirical studies of which organisational controls actually change outcomes under commercial pressure.
Participants emphasised that research should not focus only on spectacular failure cases. The more likely harm is subtle: a slow drift toward decision-making based on plausible but wrong synthetic narratives, compounded over time. Addressing this requires measurement, transparency, and cultural norms that treat validation not as a tax but as core to professional legitimacy.
17. Conclusion
The symposium’s primary contribution was to shift the ethical conversation from slogans to systems. Participants treated synthetic market research as neither inherently unethical nor inherently safe. Instead, it was discussed as a powerful methodological capability that demands a mature discipline: clear taxonomy, explicit claims, privacy testing, fairness evaluation, robust governance, and honest disclosure.
While disagreements remained, they largely reflected a healthy field in formation. The most constructive tension concerned thresholds: how much validation is enough, when consent is required, and where the line between persuasion and exploitation should be drawn. The symposium’s lasting message was that these questions cannot be left to marketing language or ad hoc judgement. They require shared standards, operational controls, and an evidence-based practice of continual benchmarking against reality.
In that sense, the symposium framed an ethical imperative that is also a scientific one: synthetic market research must remain accountable to the world it claims to model. If it does, it may expand the reach and speed of insight while maintaining public trust. If it does not, it risks becoming a high-velocity generator of confident error, and the harms will be borne by both consumers and the credibility of the research profession itself.
- Ethics treated as a property of method, not a side note.
- Synthetic is not automatically private; test for leakage.
- Granularity drives risk-twins need stronger governance.
- Over-claiming is an ethical harm when it drives decisions.
- Guardrails must be operational: policies, audits, accountability.