Can Large Language Models Simulate Human Cognition Beyond Behavioral Imitation?
arXiv cs.CL / 3/31/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper investigates whether large language models can simulate aspects of human cognition rather than only imitating observable behavior, addressing limitations of existing datasets that use synthetic traces or aggregated population data.
- It introduces a benchmark based on longitudinal publication histories of 217 AI researchers, treating each author’s work as an external proxy for individual cognitive processes.
- To test whether LLMs transfer cognitive patterns, the benchmark uses a cross-domain, temporal-shift generalization setup rather than standard within-domain evaluation.
- The authors propose a multidimensional cognitive alignment metric to measure individual-level cognitive consistency and run systematic evaluations of state-of-the-art LLMs plus enhancement techniques.
- The study is positioned as an initial empirical step answering how well current LLMs simulate human cognition and how much existing methods can improve those abilities.
Related Articles

Black Hat Asia
AI Business
[D] How does distributed proof of work computing handle the coordination needs of neural network training?
Reddit r/MachineLearning

Claude Code's Entire Source Code Was Just Leaked via npm Source Maps — Here's What's Inside
Dev.to

BYOK is not just a pricing model: why it changes AI product trust
Dev.to

AI Citation Registries and Identity Persistence Across Records
Dev.to