Train Yourself as an LLM: Exploring Effects of AI Literacy on Persuasion via Role-playing LLM Training

arXiv cs.CL / 4/6/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The study introduces LLMimic, a role-play-based, interactive, gamified AI literacy tutorial designed to proactively reduce the impact of persuasive LLM outputs.
  • Participants in a 2×3 between-subjects experiment (N=274) either watched an AI history video or trained with LLMimic, then faced realistic persuasion scenarios including charity donation, malicious money solicitation, and hotel recommendations.
  • LLMimic significantly improved AI literacy (p<.001) and reduced persuasion success across scenarios (p<.05), indicating better participant resistance to persuasive tactics.
  • In the hotel recommendation scenario, the training also increased participants’ truthfulness and social responsibility (p<.01), suggesting ethical/behavioral benefits beyond mere comprehension.
  • Overall, the authors argue LLMimic could be a scalable, human-centered mitigation strategy compared with passive defenses like detectors and disclaimers.

Abstract

As large language models (LLMs) become increasingly persuasive, there is concern that people's opinions and decisions may be influenced across various contexts at scale. Prior mitigation (e.g., AI detectors and disclaimers) largely treats people as passive recipients of AI-generated information. To provide a more proactive intervention against persuasive AI, we introduce \textbf{LLMimic}, a role-play-based, interactive, gamified AI literacy tutorial, where participants assume the role of an LLM and progress through three key stages of the training pipeline (pretraining, SFT, and RLHF). We conducted a 2 \times 3 between-subjects study (N = 274) where participants either (1) watched an AI history video (control) or (2) interacted with LLMimic (treatment), and then engaged in one of three realistic AI persuasion scenarios: (a) charity donation persuasion, (b) malicious money solicitation, or (c) hotel recommendation. Our results show that LLMimic significantly improved participants' AI literacy (p < .001), reduced persuasion success across scenarios (p < .05), and enhanced truthfulness and social responsibility levels (p<0.01) in the hotel scenario. These findings suggest that LLMimic offers a scalable, human-centered approach to improving AI literacy and supporting more informed interactions with persuasive AI.