Synthetic market research is moving fast, and vendors can look surprisingly similar in a demo because almost everyone can generate fluent, human-sounding answers.
The trick is not getting wowed by output quality - it’s figuring out whether the system is credible, repeatable, comparable, and safe to use for the decision you’re making.
This is a plain English guide for how to select a synthetic market research vendor. For more information you can also see the SMRA Vendor Evaluation Checklist (web version) and the procurement-ready PDF (download here).
If you’re still picking a shortlist, the SMRA tool recommendations post is also a useful starting point: Our Top 4 picks for Synthetic Market Research Tools.
The 60-second version
- Decide what you’re actually buying (a prompt persona? a synthetic panel? digital twins? a managed service?).
- Run “Gate 0” before the demo: ask for a minimum disclosure pack in writing.
- Do a structured demo: don’t let it turn into improvisational prompting.
- Run a small pilot that forces repeatability, sensitivity checks, and at least one benchmark.
- Score vendors the same way so the loudest demo doesn’t win by default.
If you only remember one principle from the SMRA checklist, make it this: convincing output is not validation.
Step 1: Start with the “what are we buying?” question
The fastest path to disappointment is buying “synthetic research” without pinning down what kind of product it is. A “persona” can mean anything from a single GPT prompt to a persistent agent with memory and state. The SMRA checklist calls this out directly: Category clarity: what exactly is being sold?
Ask these three questions right away
- Is this a synthetic panel, a persona generator, “digital twins”, or a research workflow tool?
- What is simulated vs measured? (and what’s derived from real respondent data, if anything?)
- What claims do you NOT make? (e.g., “replacement for fieldwork”, “represents real individuals”)
If a vendor can’t answer those clearly, it’s not necessarily a “bad” tool - but it’s a signal you should treat outputs as exploratory/hypothesis-generating rather than decision-grade.
Step 2: Run “Gate 0” (before the demo)
Gate 0 is the SMRA checklist’s way of stopping “demo theatre”. Before anyone shares screens, ask for the minimum disclosure pack in writing: Jump to Gate 0 (or use the PDF version: Vendor Evaluation Checklist (PDF)).
Gate 0: the minimum you should request
- Population frame statement: who results represent (geo/language/time window) and who is excluded.
- Study disclosure example: a completed “study disclosure label” (or equivalent), even if redacted.
- Grounding & provenance summary: what data sources are used and what is retained/deleted.
- Validation pack: test–retest stability, sensitivity testing, and at least one benchmark (or a pilot plan).
- Auditability: what is logged and what customers can export.
- Security & privacy controls: isolation between customers, retention rules, incident response posture.
- Use policy: prohibited/restricted use cases and how enforcement works.
If they can’t provide this pack, you don’t have enough information to evaluate reliability, privacy risk, or comparability. That’s exactly why Gate 0 exists.
Step 3: Make sure you’re comparing “personas” properly
Persona ambiguity is one of the biggest sources of mis-selling in this category. The SMRA checklist explicitly recommends requiring a written persona specification and using a declared capability level for comparability: Personas, twins, and comparability.
A useful shorthand is the Synthetic Persona Level (SPL) ladder referenced in the checklist: The Ten Levels of Synthetic Personas (SPL). You don’t need to “endorse” the ladder - you use it to force clarity.
Ask vendors to specify (in writing)
- Is the persona just a prompt? Or a structured profile? Or a calibrated distribution?
- Does it have memory? If yes: how is memory stored, bounded, and decayed (to avoid “perfect recall”)?
- Does it have state? (goals, beliefs, constraints, mood) and can that state be audited?
- Does “today” exist? (context streaming) - and how is provenance controlled?
- Is it connected to the world? (web/tools/APIs) - and what is logged?
- Is it single-turn Q&A, multi-turn interview, agentic workflow, or multi-agent simulation?
- What SPL level do you claim? (If they claim one, ask them to prove it.)
In practice: if Vendor A is selling prompt-only personas and Vendor B is selling persistent memory/stateful agents, those are different products - and should not be compared as if they’re equivalent.
Step 4: Don’t skip population framing (this is where most “accuracy” debates go wrong)
Synthetic market research only makes sense if it’s clear who the results represent. The SMRA checklist focuses heavily on population framing and coverage limits: Population framing & coverage.
Simple questions that prevent big mistakes
- What is the target population (country/region, language, time window)?
- How are segments defined, and what evidence supports those definitions?
- Where is it reliable vs out-of-domain?
- Do personas/panels drift over time? What is held constant vs updated?
If a vendor claims “global consumers” but can’t provide region/language-specific validation, treat it as a red flag.
Step 5: Validation that actually matters (not “it sounds real”)
The SMRA checklist is blunt: the main ethical failure mode is over-claiming - treating simulation as measurement. See: Validation: can they prove reliability?
What you should ask for
- Test–retest stability: same stimuli, same wording, same settings - how much does it vary?
- Sensitivity tests: does the conclusion flip if you tweak question wording or context slightly?
- External benchmarks: public stats, known-truth tasks, or limited fieldwork comparisons.
- Known failure modes: where does it break, and what guardrails exist?
Easy red flags
- Validation framed as “look how human it sounds.”
- No willingness to benchmark against even a small real sample where feasible.
- Only screenshots, no methodology you can replicate.
Step 6: Run a small pilot (this catches 80% of problems)
The checklist recommends a minimal pilot that forces repeatability and benchmarking: Pilot plan: the minimum test you should run.
A simple pilot you can run in 1–2 weeks
- Pick two study types you actually use (e.g., message test + concept test).
- Lock the protocol: fixed stimuli, fixed wording, fixed run settings.
- Run test–retest: run the same study multiple times; record variance and ask why it varies.
- Run sensitivity checks: small wording tweaks; see if conclusions change.
- Run at least one benchmark: a small human sample, or an external known reference.
- Require disclosures on every output: population frame, grounding class, limitations, reproducibility notes.
If the vendor can’t produce stable directional conclusions under repeat runs and can’t show at least one credible benchmark alignment, you’ve learned something valuable: the tool may still be useful for ideation, but it’s not decision-grade for your use case.
Step 7: Score vendors consistently (so the best demo doesn’t win)
The SMRA checklist includes a simple 0–2 scoring rubric across categories like population frame, persona specification, grounding/provenance, validation/benchmarking, disclosure, auditability, privacy/security, and misuse safeguards: Scoring rubric.
You don’t need a complicated model. The goal is consistency: you want a paper trail that explains why Vendor X was chosen and what limitations you accepted.
How to build your shortlist (without overthinking it)
If you’re still early and you’re not sure which tools to test, start with the SMRA scenario-based recommendations: Our Top 4 picks for Synthetic Market Research Tools. It frames tools by what you’re trying to do (e.g., cheap content optimization vs recreating your customers vs managed GTM deliverables vs high-stakes enterprise use).
A nice practical workflow is:
- Pick 2–4 vendors that match your scenario.
- Apply Gate 0 to all of them.
- Only demo the ones that can provide the minimum disclosure pack.
- Pilot the top 1–2 with the same protocol and score them with the same rubric.
Extra: Helpful SMRA pages to keep open while evaluating
- Quick vendor recommendations
- Vendor Evaluation Checklist (web)
- Vendor Evaluation Checklist (PDF)
- SMRA Resources & Templates hub
- Standards baseline (disclosure, validation, auditability, privacy posture)
- SPL ladder for persona comparison
Final note: be friendly, but be stubborn
Most vendors are not acting in bad faith - they’re operating in a young category where terminology is messy and evaluation norms are still forming. Your job as a buyer is to make the evaluation concrete: define the population, define the persona, demand evidence, run a pilot, and document limitations. If you do that, you’ll end up with tools that are genuinely useful - and you’ll avoid the classic “amazing demo, disappointing reality” trap.
- Decide what you’re buying: prompt personas, panels, twins, or a managed service.
- Run Gate 0 before demos; treat “no disclosure pack” as exploratory-only.
- Force repeatability: structured demos, pilots with stability/sensitivity tests, and at least one benchmark.
- Compare personas fairly: memory / state / grounding / connectivity claims must be written and testable.
- Score vendors consistently so the flashiest demo doesn’t win by default.