An Interactive LLM-Based Simulator for Dementia-Related Activities of Daily Living

arXiv cs.RO / 4/1/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisTools & Practical UsageModels & Research

Key Points

  • The paper introduces a web-based, interactive LLM simulator (using gpt-5-mini) that generates multi-turn, dementia-severity- and care-setting-conditioned patient behaviors during ADL assistance scenarios, paired with lightweight behavioral cues.
  • Users can configure dementia severity, care setting, and the specific ADL, then act as caregivers via free-text responses or strategy-scaffolded suggestions (Recognition, Negotiation, Facilitation, Validation), while rating the realism of each simulated patient turn.
  • An expert-in-the-loop formative study with 14 dementia-care experts (18 sessions, 112 rated turns) found the simulated behaviors were judged moderately to highly plausible, with an average session of about six turns.
  • Experts frequently authored custom caregiver replies (54.5%), and the most used strategies were Recognition and Facilitation, reflecting which interaction patterns best resonated in the simulated ADL contexts.
  • Critiques were analyzed into a six-category failure-mode taxonomy, highlighting recurring issues with ADL grounding and care-setting consistency that inform prompt/workflow refinements and future evidence-driven co-simulation for caregiver training and assistive AI/robot policy development.

Abstract

Effective dementia caregiving requires training and adaptive communication, but assistive AI and robotics are constrained by a lack of context-rich, privacy-sensitive data on how people living with Alzheimer's disease and related dementias (ADRD) behave during activities of daily living (ADLs). We introduce a web-based simulator that uses a large language model (gpt-5-mini) to generate multi-turn, severity- and care-setting-conditioned patient behaviors during ADL assistance, pairing utterances with lightweight behavioral cues (in parentheses). Users set dementia severity, care setting (and time in setting), and ADL; after each patient turn they rate realism (1-5) with optional critique, then respond as the caregiver via free text or by selecting/editing one of four strategy-scaffolded suggestions (Recognition, Negotiation, Facilitation, Validation). We ran an online formative expert-in-the-loop study (14 dementia-care experts, 18 sessions, 112 rated turns). Simulated behavior was judged moderately to highly plausible, with a typical session length of six turns. Experts wrote custom replies for 54.5 percent of turns; Recognition and Facilitation were the most-used suggested strategies. Thematic analysis of critiques produced a six-category failure-mode taxonomy, revealing recurring breakdowns in ADL grounding and care-setting consistency and guiding prompt/workflow refinements. The simulator and logged interactions enable an evidence-driven refinement loop toward validated patient-caregiver co-simulation and support data collection, caregiver training, and assistive AI and robot policy development.