How Frontier LLMs Adapt to Neurodivergence Context: A Measurement Framework for Surface vs. Structural Change in System-Prompted Responses

arXiv cs.AI / 5/4/2026

💬 OpinionSignals & Early TrendsTools & Practical UsageModels & Research

Key Points

  • The paper studies whether frontier chat-based LLMs change their responses when neurodivergence (ND) context is provided in system prompts, and characterizes what kinds of changes occur.
  • It introduces NDBench, a publicly released benchmark with 576 outputs across two frontier models, multiple system-prompt variants, four ND profiles, and 24 prompts (including an adversarial masking strategy).
  • The authors find consistent evidence of ND-related adaptation: fully instructed prompt conditions produce longer, more structured responses with more headings and more granular step-by-step detail.
  • They conclude the adaptation is mainly structural rather than lexical in nature, since list density changes little while heading frequency and per-step detail increase.
  • ND persona assertion alone does not reliably reduce harmful tendencies; only explicitly instructed settings show substantial decreases, and harm-assessment reliability varies by dimension (masking/reinforcement and validation quality outperform others).

Abstract

We examine if frontier chat-based large language models (LLMs) adjust their outputs based on neurodivergence (ND) context in system prompts and describe the nature of these adjustments. Specifically, we propose NDBench, a 576-output benchmark involving two frontier models, three system prompt types (baseline, ND-profile assertion, and ND-profile assertion with explicit instructions for adjustments), four canonical ND profiles, and 24 prompts across four categories, one of which involves an adversarial masking strategy. Four trends emerge consistently from our findings. First, LLMs show significant adaptation under ND context, where fully instructed conditions yield lengthier and more structured outputs, characterized by higher token counts, more headings, and more granular steps (p < 10^-8, Holm-corrected). Second, such adaptation is largely structural in nature: although list density does not change much, there is a marked rise in the frequency of headings and per-step detail. Third, ND persona assertion alone fails to suppress potentially harmful tendencies, as masking-reinforcement decreases only in explicitly instructed cases (36-44% reduction); the reduction rate barely changes in persona assertion conditions. Moreover, reliability analysis of LLM-based harm assessment reveals that only two out of the six dimensions (masking and reinforcement, validation quality) exceed the pre-defined inter-judge agreement criterion (alpha >= 0.67) and thus can be considered primary results. NDBench is made publicly available along with its prompts, outputs, code, and other resources, forming a reproducible framework for auditing future LLMs' adaptation to ND awareness.