LLLMs: A Data-Driven Survey of Evolving Research on Limitations of Large Language Models
arXiv cs.CL / 3/13/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The survey uses a data-driven, semi-automated approach to review limitations of LLMs (LLLMs) from 2022 to early 2025, analyzing a corpus of 250,000 ACL and arXiv papers with keyword filtering, LLM-based classification, expert validation, and topic clustering (HDBSCAN+BERTopic and LlooM).
- It reports that the share of LLM-related papers has grown fivefold in ACL and eightfold in arXiv since 2022, with LLLMs making up over 30% of LLM papers by 2025.
- Reasoning is the most studied limitation, followed by generalization, hallucination, bias, and security, and the arXiv dataset shows shifting emphasis toward security risks, alignment, hallucinations, knowledge editing, and multimodality.
- The authors release a dataset of annotated abstracts and a validated methodology publicly on GitHub, enabling reproducibility and further research.




