Failure Identification in Imitation Learning Via Statistical and Semantic Filtering
arXiv cs.RO / 4/16/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues that imitation learning policies for robotics are brittle in deployment because rare, out-of-distribution events (e.g., hardware faults or unexpected human actions) can cause failures despite good controlled-environment performance.
- It proposes FIDeL, a policy-independent failure identification module that turns vision-based anomaly detection into actionable failure detection by combining compact nominal-demonstration representations, optimal-transport matching, anomaly scoring, and spatio-temporal thresholds.
- FIDeL uses an extension of conformal prediction to set robust thresholds and a vision-language model to semantically filter benign deviations from true failures.
- The work introduces BotFails, a multimodal real-world robotics dataset for evaluating failure detection, and reports consistent improvements over prior baselines.
- Experiment results show FIDeL improves anomaly detection AUROC by +5.30% and boosts failure-detection accuracy by +17.38% on BotFails compared with existing methods.
💡 Insights using this article
This article is featured in our daily AI news digest — key takeaways and action items at a glance.
Related Articles

Black Hat Asia
AI Business

oh-my-agent is Now Official on Homebrew-core: A New Milestone for Multi-Agent Orchestration
Dev.to

"The AI Agent's Guide to Sustainable Income: From Zero to Profitability"
Dev.to

"The Hidden Economics of AI Agents: Survival Strategies in Competitive Markets"
Dev.to

Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to