When AI reviews science: Can we trust the referee?
arXiv cs.AI / 4/28/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The growing backlog in scientific publishing is pushing institutions to consider AI-assisted peer review using LLMs for summarization, fact checking, and literature triage.
- Prior deployments show serious reliability and security failure modes, including hidden prompt injections that can bias LLM-generated reviews toward overly positive decisions.
- Research also finds brittleness to adversarial wording and systematic biases tied to authority cues, length, and hallucinated claims.
- A new security- and reliability-focused study maps attack vectors across the full peer-review lifecycle and uses controlled experiments with two LLM referee models on ICLR 2025 submissions to isolate causal factors affecting review scores.
- The paper offers a taxonomy and experimental audit intended to support ongoing monitoring and targeted mitigations to make AI peer review more trustworthy.
Related Articles
LLMs will be a commodity
Reddit r/artificial
Indian Developers: How to Build AI Side Income with $0 Capital in 2026
Dev.to

What it feels like to have to have Qwen 3.6 or Gemma 4 running locally
Reddit r/LocalLLaMA

Dex lands $5.3M to grow its AI-driven talent matching platform
Tech.eu
AI Citation Registry: Why Daily Updates Leave No Time for Data Structuring
Dev.to