Large Language Models for Market Research: A Data-augmentation Approach

arXiv stat.ML / 4/20/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies how large language models can support market research for conjoint analysis, where gathering consumer preference data is typically costly and hard to scale.
  • It argues that simply substituting real survey responses with LLM-generated data can introduce bias and create a meaningful gap between LLM-simulated and human data.
  • The authors propose a statistical data-augmentation method that combines LLM-generated and real data to produce estimators that are consistent and asymptotically normal.
  • Experiments on COVID-19 vaccine preferences and sports car choices show substantial reductions in estimation error and reported data/cost savings of about 24.9% to 79.8%, while naive substitution approaches do not achieve similar savings.
  • Overall, the work concludes that LLM-generated data should be used as a complementary input rather than a direct replacement, but can be highly effective within the proposed rigorous framework.

Abstract

Large Language Models (LLMs) have transformed artificial intelligence by excelling in complex natural language processing tasks. Their ability to generate human-like text has opened new possibilities for market research, particularly in conjoint analysis, where understanding consumer preferences is essential but often resource-intensive. Traditional survey-based methods face limitations in scalability and cost, making LLM-generated data a promising alternative. However, while LLMs have the potential to simulate real consumer behavior, recent studies highlight a significant gap between LLM-generated and human data, with biases introduced when substituting between the two. In this paper, we address this gap by proposing a novel statistical data augmentation approach that efficiently integrates LLM-generated data with real data in conjoint analysis. This results in statistically robust estimators with consistent and asymptotically normal properties, in contrast to naive approaches that simply substitute human data with LLM-generated data, which can exacerbate bias. We further present a finite-sample performance bound on the estimation error. We validate our framework through an empirical study on COVID-19 vaccine preferences, demonstrating its superior ability to reduce estimation error and save data and costs by 24.9% to 79.8%. In contrast, naive approaches fail to save data due to the inherent biases in LLM-generated data compared to human data. Another empirical study on sports car choices validates the robustness of our results. Our findings suggest that while LLM-generated data is not a direct substitute for human responses, it can serve as a valuable complement when used within a robust statistical framework.