Skip to content
Article Nielsen Norman Group Aug 2025

AI-simulated behavior in user research — NN/g evaluation

What the article covers

Nielsen Norman Group examines three academic studies that test whether AI-powered digital twins and synthetic users can replicate real human responses in research settings. The studies cover survey-based digital twins, interview-based digital twins, and synthetic users built to represent population-level groups.

Context

As AI user simulation tools proliferate, research teams face pressure to adopt them for speed and cost savings. This article provides a rigorous evaluation of what the current evidence actually shows, drawing from peer-reviewed research rather than vendor claims.

Key takeaway

The findings are more nuanced than either advocates or skeptics typically present. Digital twins built on extensive interview data can achieve remarkable individual-level and population-level predictions. The simplest technique, augmenting LLM prompts with interview transcripts, produces better results than more complex approaches. Synthetic users built without individual data can predict population-level trends but generate less variability than real humans, and accuracy depends heavily on demographic group, task, and context. The practical recommendation: start with data-rich touchpoints you already collect and experiment with prompt-augmented digital twins before investing in more elaborate approaches. The article is also clear about ethical concerns including misrepresentation, bias, and loss of agency for simulated individuals.

Who should read this

Research leaders evaluating synthetic user tools, academics studying AI simulation methods, and practitioners who need evidence-based arguments for or against incorporating AI-simulated participants into their research practice.