LLLMs: A Data-Driven Survey of Evolving Research on Limitations of Large Language Models
arXiv cs.CL / 3/13/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The survey uses a data-driven, semi-automated approach to review limitations of LLMs (LLLMs) from 2022 to early 2025, analyzing a corpus of 250,000 ACL and arXiv papers with keyword filtering, LLM-based classification, expert validation, and topic clustering (HDBSCAN+BERTopic and LlooM).
- It reports that the share of LLM-related papers has grown fivefold in ACL and eightfold in arXiv since 2022, with LLLMs making up over 30% of LLM papers by 2025.
- Reasoning is the most studied limitation, followed by generalization, hallucination, bias, and security, and the arXiv dataset shows shifting emphasis toward security risks, alignment, hallucinations, knowledge editing, and multimodality.
- The authors release a dataset of annotated abstracts and a validated methodology publicly on GitHub, enabling reproducibility and further research.
Related Articles
Is AI becoming a bubble, and could it end like the dot-com crash?
Reddit r/artificial

Externalizing State
Dev.to

I made a 'benchmark' where LLMs write code controlling units in a 1v1 RTS game.
Dev.to

My AI Does Not Have a Clock
Dev.to

From Early Adopter to AI Instructor: Teaching 500 Engineers to Build with LLMs
Dev.to