The Format Tax
arXiv cs.CL / 4/7/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues that asking an LLM to output structured formats (JSON/XML/LaTeX/Markdown) should not reduce reasoning quality, but structured-output requirements significantly degrade performance in open-weight models.
- The study finds that constrained decoding explains only a small part of the loss; most accuracy degradation comes directly from the prompt-level instructions demanding a specific format.
- Across six open-weight models (and additional comparisons with closed-weight models), the authors show that separating/decoupling reasoning from formatting substantially recovers lost accuracy.
- The proposed recovery strategies include generating freeform first and then reformatting in a second pass, or allowing extended “thinking” within a single generation before producing the final structured output.
- Closed-weight models are observed to exhibit little to no “format tax,” implying the issue is largely a current open-weight-model capability gap rather than a fundamental limitation of structured generation.
Related Articles
[R] The ECIH: Model Modeling Agentic Identity as an Emergent Relational State [R]
Reddit r/MachineLearning
Google DeepMind Unveils Project Genie: The Dawn of Infinite AI-Generated Game Worlds
Dev.to
Artificial Intelligence and Life in 2030: The One Hundred Year Study onArtificial Intelligence
Dev.to
From Booth Chaos to Scalable Conversations: AI for Hyper-Personalized Follow-Up
Dev.to
AI in 2030: 20 Powerful Trends That Will Shape the Future
Dev.to