AI and My Values: User Perceptions of LLMs' Ability to Extract, Embody, and Explain Human Values from Casual Conversations

arXiv cs.CL / 3/30/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces VAPT (Value-Alignment Perception Toolkit) to pragmatically study whether LLMs can extract, embody, and explain users’ human values based on casual conversations.
  • In a month-long user texting study with 20 participants, the toolkit assessed the alignment between users’ perceived values and the AI’s actions and justifications.
  • Thirteen of the participants reported being convinced that AI can understand human values, indicating strong perceived value-reflection despite uncertainty about actual understanding.
  • The authors caution about a potential “weaponized empathy” pattern in which value-aware conversational agents may still be welfare-misaligned.
  • VAPT is proposed as an evaluation method plus a set of design implications—aiming to support more transparent, safeguarded value-alignment as LLM capabilities become harder to interpret and more ubiquitous.

Abstract

Does AI understand human values? While this remains an open philosophical question, we take a pragmatic stance by introducing VAPT, the Value-Alignment Perception Toolkit, for studying how LLMs reflect people's values and how people judge those reflections. 20 participants texted a chatbot over a month, then completed a 2-hour interview with our toolkit evaluating AI's ability to extract (pull details regarding), embody (make decisions guided by), and explain (provide proof of) their values. 13 participants ultimately left our study convinced that AI can understand human values. Thus, we warn about "weaponized empathy": a design pattern that may arise in interactions with value-aware, yet welfare-misaligned conversational agents. VAPT offers a new way to evaluate value-alignment in AI systems. We also offer design implications to evaluate and responsibly build AI systems with transparency and safeguards as AI capabilities grow more inscrutable, ubiquitous, and posthuman into the future.