How Open Must Language Models be to Enable Reliable Scientific Inference?
arXiv cs.CL / 3/30/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper examines how the openness or closedness of language models affects the reliability of scientific inferences derived from research using those models.
- It argues that restrictions on information about model construction and deployment can introduce threats to scientific inference, making many closed models poorly suited for scientific applications.
- The authors note exceptions where some closed models may still support scientific purposes, but they generally emphasize the risk of unverifiable or non-reproducible behavior.
- They propose mitigation approaches and recommend that researchers systematically identify inference threats, document mitigation steps, and provide explicit justifications for choosing a specific model.
Related Articles

Mr. Chatterbox is a (weak) Victorian-era ethically trained model you can run on your own computer
Simon Willison's Blog
Beyond the Chatbot: Engineering Multi-Agent Ecosystems in 2026
Dev.to

I missed the "fun" part in software development
Dev.to

The Billion Dollar Tax on AI Agents
Dev.to

Hermes Agent: A Self-Improving AI Agent That Runs Anywhere
Dev.to