Sample Transform Cost-Based Training-Free Hallucination Detector for Large Language Models
arXiv cs.AI / 3/25/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper proposes a training-free hallucination detector for large language models by using distribution complexity inferred from prompt-conditioned responses.
- It computes Wasserstein (optimal-transport) distances between sets of token embeddings from pairwise samples to build a Wasserstein distance matrix that reflects transformation costs.
- The authors derive two complementary signals—AvgWD (average transformation cost) and EigenWD (cost-complexity via eigen-structure)—to quantify likelihood of hallucination.
- The method is extended to black-box LLM settings using a “teacher forcing” approach with an accessible teacher model.
- Experiments on multiple models and datasets show AvgWD and EigenWD are competitive with strong uncertainty baselines and exhibit complementary behaviors, supporting “distribution complexity” as a truthfulness signal.
Related Articles
Santa Augmentcode Intent Ep.6
Dev.to

Your Agent Hired Another Agent. The Output Was Garbage. The Money's Gone.
Dev.to
ClawRouter vs TeamoRouter: one requires a crypto wallet, one doesn't
Dev.to
Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to

Palantir’s billionaire CEO says only two kinds of people will succeed in the AI era: trade workers — ‘or you’re neurodivergent’
Reddit r/artificial