How Frontier LLMs Adapt to Neurodivergence Context: A Measurement Framework for Surface vs. Structural Change in System-Prompted Responses
arXiv cs.AI / 5/4/2026
💬 OpinionSignals & Early TrendsTools & Practical UsageModels & Research
Key Points
- The paper studies whether frontier chat-based LLMs change their responses when neurodivergence (ND) context is provided in system prompts, and characterizes what kinds of changes occur.
- It introduces NDBench, a publicly released benchmark with 576 outputs across two frontier models, multiple system-prompt variants, four ND profiles, and 24 prompts (including an adversarial masking strategy).
- The authors find consistent evidence of ND-related adaptation: fully instructed prompt conditions produce longer, more structured responses with more headings and more granular step-by-step detail.
- They conclude the adaptation is mainly structural rather than lexical in nature, since list density changes little while heading frequency and per-step detail increase.
- ND persona assertion alone does not reliably reduce harmful tendencies; only explicitly instructed settings show substantial decreases, and harm-assessment reliability varies by dimension (masking/reinforcement and validation quality outperform others).
Related Articles

Black Hat USA
AI Business
A very basic litmus test for LLMs "ok give me a python program that reads my c: and put names and folders in a sorted list from biggest to small"
Reddit r/LocalLLaMA

ALM on Power Platform: ADO + GitHub, the best of both worlds
Dev.to

Iron Will, Iron Problems: Kiwi-chan's Mining Misadventures! 🥝⛏️
Dev.to
Experiment: Does repeated usage influence ChatGPT 5.4 outputs in a RAG-like setup?
Dev.to