Banana100: Breaking NR-IQA Metrics by 100 Iterative Image Replications with Nano Banana Pro
arXiv cs.CV / 4/7/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper identifies a key failure mode in multi-turn, multi-modal image editing: repeated edits cause iterative degradation that accumulates visible noise and can break adherence to even simple instructions.
- It introduces Banana100, a new dataset of 28,000 images created via 100 iterative editing steps across varied textures and image content, specifically targeting this degradation phenomenon.
- The authors report that common no-reference image quality assessment (NR-IQA) metrics fail to reliably flag heavily degraded images, with none of 21 popular metrics consistently scoring degraded images lower than clean ones.
- The dual breakdown—both in image generators and in evaluators—raises concerns about training stability and the safety of deployed agentic systems if low-quality synthetic outputs bypass quality filters.
- The authors release code and data to support building more robust editing systems and more reliable quality evaluation for multi-turn agentic workflows.
💡 Insights using this article
This article is featured in our daily AI news digest — key takeaways and action items at a glance.
Related Articles

Black Hat Asia
AI Business
[R] The ECIH: Model Modeling Agentic Identity as an Emergent Relational State [R]
Reddit r/MachineLearning
Google DeepMind Unveils Project Genie: The Dawn of Infinite AI-Generated Game Worlds
Dev.to
Artificial Intelligence and Life in 2030: The One Hundred Year Study onArtificial Intelligence
Dev.to
Stop waiting for Java to rebuild! AI IDEs + Zero-Latency Hot Reload = Magic
Dev.to