Large Language Models for Missing Data Imputation: Understanding Behavior, Hallucination Effects, and Control Mechanisms
arXiv cs.AI / 3/25/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper presents a large-scale benchmarking study of five LLMs for tabular missing-data imputation using zero-shot prompt engineering and comparing them against six state-of-the-art traditional imputation baselines.
- Evaluations span 29 datasets (including nine synthetic sets) across missingness mechanisms MCAR, MAR, and MNAR and missing rates up to 20%, enabling more systematic cross-method comparisons than prior work.
- Results show LLMs—especially Gemini 3.0 Flash and Claude 4.5 Sonnet—typically outperform traditional methods on real-world open-source datasets.
- The study finds the LLM advantage is likely linked to pretraining-induced familiarity with domain-specific patterns, while traditional methods like MICE outperform LLMs on synthetic datasets, indicating LLMs rely more on semantic context than statistical reconstruction.
- A key practical trade-off is identified: LLM-based imputation achieves higher quality but requires substantially greater computational time and monetary cost than classical approaches.
Related Articles
Santa Augmentcode Intent Ep.6
Dev.to

Your Agent Hired Another Agent. The Output Was Garbage. The Money's Gone.
Dev.to
ClawRouter vs TeamoRouter: one requires a crypto wallet, one doesn't
Dev.to
Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to

Palantir’s billionaire CEO says only two kinds of people will succeed in the AI era: trade workers — ‘or you’re neurodivergent’
Reddit r/artificial