User Research

Synthetic users for research: what they are and where they fall short

The pitch for synthetic users is appealing: instant research subjects, any profile, no scheduling required. The problem is that the answers are not real. They are statistically plausible text generated by a language model, not observations of actual human behavior.

CleverX Team ·
Synthetic users for research: what they are and where they fall short

The pitch for synthetic users is straightforward and genuinely appealing: instant research subjects available on demand, any profile, any demographic, no scheduling required. A team that needs to understand how a 45-year-old logistics manager would react to a new software workflow can generate a synthetic version of that person and ask them questions right now. No recruitment lead time. No incentive budget. No participant no-shows. Just answers.

The problem is that the answers are not real. They are statistically plausible text generated by a language model based on patterns in its training data. The logistics manager responding to your questions does not exist. The responses reflect what a language model predicts someone matching that description might say, not what a real person in that role actually thinks, does, or experiences. For teams making product decisions, that distinction matters enormously.

Synthetic user tools have entered the research market at an interesting moment. AI interview tools that conduct real research with real participants have also become more capable and accessible, which means the speed argument that makes synthetic users superficially attractive is weaker than it used to be. Understanding exactly what synthetic users can and cannot do, and where real AI-assisted research with actual participants is now fast enough to make synthetic shortcuts unnecessary, is the relevant question for research teams in 2026.

What synthetic user tools actually do

Synthetic user tools use large language models to generate responses that simulate how a hypothetical user with specified characteristics might react to interview questions, product concepts, survey prompts, or design stimuli. A researcher describes a user profile, the tool instantiates a language model persona matching that profile, and the model generates responses as if it were that person.

Some tools accept a structured persona description and generate interview transcripts in the voice of that persona. Others allow researchers to define user segments and simulate survey responses or usability session behaviors across those segments. The output is AI-generated text that statistically resembles what a person matching the specified profile might say, based on the distribution of text in the language model’s training data that correlates with people like the described user.

This is the critical technical point: synthetic user responses are predictions based on text patterns, not observations of human behavior. Language models learn from what people have written about themselves and their experiences online, in published research, in product reviews, and in every other form of text that entered their training data. When a synthetic user tool generates a response from a “senior IT manager with 15 years of enterprise software experience,” it is generating text that fits the statistical patterns associated with how people matching that description write and speak in the training corpus. It is not accessing any actual IT manager’s actual experience with anything.

Where synthetic users have legitimate uses

The legitimate applications for synthetic users are narrower than the marketing suggests, but they are real.

Early-stage assumption pressure-testing is the strongest use case. Before a prototype exists, when a product team is working through the plausibility of a proposed concept or trying to anticipate where user objections might arise, running the concept description through a synthetic user conversation can surface potential friction points and challenge internal assumptions cheaply. The bar for this kind of input is low: the team is looking for directional prompts to think harder about, not validated findings to act on. Synthetic users can generate useful provocations at this stage without claiming to represent real user behavior.

Discussion guide stress-testing is a second legitimate application. Running a draft discussion guide through a synthetic user interaction reveals confusing questions, leading question structures, logical gaps in the guide flow, and ambiguous phrasing before exposing real participants to a guide that has not been properly reviewed. This is using synthetic users as a proxy for a colleague reading the guide critically, not as a research data source. The simulation does not need to be accurate about real user behavior to serve this function.

Researcher training and practice is a third application where synthetic accuracy does not matter. New researchers learning to moderate sessions need practice environments that produce plausible-sounding responses to practice probing and follow-up skills against. Synthetic users provide this without consuming real participant time or budget on training exercises.

Hypothesis generation before real research rounds out the appropriate use cases. Synthetic user responses can surface potential themes and lines of inquiry that researchers had not considered, which can improve the quality of discussion guides and analytical frameworks before the real research begins. Using synthetic output as a hypothesis generation tool rather than a finding generation tool keeps the synthetic layer in its appropriate role: prompting better questions for real participants to answer, not substituting for asking real participants.

Where synthetic users fundamentally fail

The failures are not edge cases. They apply to every core research question that product and design decisions actually depend on.

Synthetic users cannot produce genuine behavioral data. Real usability research observes what users actually do: the navigation paths that no designer anticipated, the workarounds users invented because the intended path did not work for them, the moments where users abandon a task mid-flow because their mental model diverged from the design’s assumption. Language models generate responses based on statistical patterns in text, not based on actual human behavior. There is no version of synthetic user simulation that can produce the genuinely unexpected finding, because LLMs can only generate text that statistically fits their training distribution. Real research regularly surfaces findings that nobody predicted. Synthetic research structurally cannot.

Synthetic users reflect LLM training biases rather than real user populations. Training data does not uniformly represent all users. Underrepresented groups, non-Western markets, elderly users, users with disabilities, users with low digital literacy, and users from economic contexts that generate less online text are all represented less accurately in synthetic simulations than in real participant research. For research on consumer products targeting mainstream demographics in markets well-represented in LLM training data, synthetic responses are at least partially grounded in real patterns. For research on specialized professional roles, non-English-speaking markets, or populations outside the comfortable center of internet demographics, synthetic simulations are generating statistically plausible fictions that have no reliable grounding in the actual population the product is designed for.

Synthetic users cannot capture embodied and contextual experience. A user completing a checkout flow while distracted, on a phone, during a commute, with a slow connection, after a frustrating prior experience with a different app is fundamentally different from a synthetic simulation responding to abstract questions about checkout preferences. Physical environment, cognitive load, emotional state, and situational constraints shape real user behavior in ways that no textual persona description can encode. The contextual grounding that makes user research valuable is precisely what synthetic simulation strips away.

Synthetic users generate responses that are consistent with their persona description by construction. Real users are inconsistent. They say one thing and do another. They express a preference that contradicts their revealed behavior. They describe their workflow inaccurately because they have never been asked to observe themselves doing it. These inconsistencies are not noise to be eliminated; they are often the most analytically significant findings in a research study. Synthetic simulations, optimized to generate persona-consistent responses, cannot produce the contradictions and surprises that real participants reliably provide.

The false confidence problem

The most serious risk of synthetic user research is not that it produces wrong answers. It is that it produces answers with a confidence-projecting format that makes wrong answers look like findings.

A synthetic user interview transcript looks like a real interview transcript. It has participant responses organized by moderator question. It contains quotes that could plausibly appear in a real research report. A team that runs ten synthetic user sessions and reviews the resulting transcripts receives output that strongly resembles the deliverable of a real research study. The temptation to treat it as such is predictable.

The difference is that a real research transcript reflects a real person’s actual thoughts and behaviors, however imperfectly captured. A synthetic transcript reflects a language model’s prediction of what a text-matching-that-profile would produce, which is a different thing entirely. Acting on synthetic user findings as if they were real research findings is not using uncertain data; it is using data that is structurally incapable of grounding the decisions it is being used to justify.

Teams that make product decisions based on synthetic user data are effectively making decisions based on AI-generated consensus opinions about what users are probably like, which is a sophisticated form of confirmation bias. The output tells the team what it seems like users would think, filtered through a model’s training distribution, which correlates more strongly with existing consensus assumptions about users than with any actual user population. It can make wrong assumptions feel validated rather than challenged.

Why the speed argument no longer holds

The most common justification for synthetic users is speed. Real participant research takes time to recruit, schedule, and run. Synthetic users are available immediately. For teams that need a signal today, not in ten days, synthetic users seem to solve a real operational problem.

This argument has weakened substantially as AI-assisted real participant research has matured. CleverX’s AI Interview Agent conducts asynchronous interviews with verified professionals from a pool of 8 million participants across 150 or more countries. A research team that needs to understand how enterprise IT managers respond to a new infrastructure tool can deploy an AI Interview Agent session to qualified, verified participants today and have real transcripts to analyze by tomorrow morning. The recruitment, scheduling, and session facilitation steps that historically extended research timelines by ten to fourteen days have been compressed to same-day or next-day turnaround without sacrificing the fundamental quality of real participant data.

At one dollar per credit, the cost of running real participant research through CleverX for a quick directional study is low enough that the cost argument for synthetic substitutes is similarly weakened. Running ten AI-moderated sessions with real professionals who match the target profile costs a fraction of what a traditional recruitment and moderation cycle would cost and produces findings grounded in real behavior rather than model predictions. See what are AI-moderated interviews for how AI moderation with real participants works, and AI usability testing tools for AI tools that augment real research without replacing real participants.

The right mental model

Synthetic users are a thinking tool, not a research tool. Using them to pressure-test assumptions, improve discussion guides, generate hypotheses, or train new researchers captures their real value without misrepresenting what they produce. Using them as a substitute for research with real participants produces findings that appear authoritative while being grounded in nothing but a language model’s statistical imagination.

Real research with real participants remains the only reliable way to understand what users actually do, think, and feel. The gap between what AI can plausibly simulate and what real human experience contains is exactly the gap that research exists to close.

Frequently asked questions

What are synthetic users in research?

Synthetic users are AI-generated simulations of user behavior and responses, created by prompting large language models to respond as if they were a person matching a specified demographic or behavioral profile. They produce interview-like transcripts, survey-like responses, or simulated usability session behavior without involving any real human participants. Their outputs are statistically plausible text generated from model training data, not observations of actual human behavior.

Can synthetic users replace real participant research?

No. Synthetic users cannot produce genuine behavioral data, cannot reflect underrepresented user populations accurately, cannot capture embodied or contextual experience, and cannot generate the unexpected findings that real research reliably produces. They are appropriate as thinking tools for assumption pressure-testing, discussion guide review, and hypothesis generation. They are not appropriate as substitutes for research with real participants in any context where findings will inform product or design decisions.

When is it appropriate to use synthetic users?

Synthetic users have legitimate applications in four specific contexts: pressure-testing product assumptions before a prototype exists, reviewing discussion guides for leading questions and structural problems, training new researchers in session moderation and analysis, and generating initial hypotheses to investigate through real research. In all four cases, the synthetic output is a prompt for better human thinking, not a research finding. The key constraint is never treating synthetic user output as data about real users.

How does AI-moderated research with real participants differ from synthetic user research?

AI-moderated research with real participants uses AI to conduct or assist in conducting interviews with actual human participants who have been recruited and verified as matching the study’s criteria. The participants are real people providing their genuine experiences and opinions. The AI handles facilitation, adaptive probing, and session logistics. Synthetic user research uses AI to simulate what a hypothetical participant might say without involving any real person. The difference is the difference between studying actual human behavior and generating a statistical prediction of what human behavior might look like.