Exploring the System 1 Thinking Capability of Large Reasoning Models
arXiv cs.CL / 5/4/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper investigates “System 1 thinking” in Large Reasoning Models (LRMs), focusing on their ability to answer intuitively and efficiently using minimal tokens.
- It introduces S1-Bench, a multi-domain and multilingual benchmark designed for model-simple System 1 questions.
- Experiments across 28 LRMs show that they tend to be both less accurate and less efficient on System 1-style problems than expected.
- The study finds that current efficient reasoning techniques may not generalize well to simple questions, and may trade off accuracy to achieve efficiency.
- The authors observe early difficulty awareness in LRMs, with lower confidence, and suggest that difficulty is implicitly represented in hidden states.
Related Articles
A very basic litmus test for LLMs "ok give me a python program that reads my c: and put names and folders in a sorted list from biggest to small"
Reddit r/LocalLLaMA

ALM on Power Platform: ADO + GitHub, the best of both worlds
Dev.to

Experiment: Does repeated usage influence ChatGPT 5.4 outputs in a RAG-like setup?
Dev.to

Find 12 high-volume, low-competition GEO content topics Topify.ai should rank on
Dev.to

When a memorized rule fits your bug too well: a meta-trap of agent workflows
Dev.to