Seeing Isn't Believing: Mitigating Belief Inertia via Active Intervention in Embodied Agents

arXiv cs.CL / 4/21/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper studies how LLM-based embodied agents can make suboptimal decisions because they fail to adequately incorporate environmental feedback that contradicts their internal beliefs.
  • It formalizes this issue as “belief inertia,” where agents continue to rely on prior beliefs even after receiving explicit observations.
  • To mitigate belief inertia, the authors propose active belief intervention via an Estimate-Verify-Update (EVU) mechanism that predicts outcomes, verifies them against observations using explicit reasoning, and updates beliefs accordingly.
  • EVU is presented as a unified, text-based belief-state generation and intervention method that can be integrated into both prompting-based and training-based agent reasoning.
  • Experiments on three embodied benchmarks show consistent improvements in task success rates, and additional analyses confirm EVU effectively reduces belief inertia.

Abstract

Recent advancements in large language models (LLMs) have enabled agents to tackle complex embodied tasks through environmental interaction. However, these agents still make suboptimal decisions and perform ineffective actions, as they often overlook critical environmental feedback that differs from their internal beliefs. Through a formal probing analysis, we characterize this as belief inertia, a phenomenon where agents stubbornly adhere to prior beliefs despite explicit observations. To address this, we advocate active belief intervention, moving from passive understanding to active management. We introduce the Estimate-Verify-Update (EVU) mechanism, which empowers agents to predict expected outcomes, verify them against observations through explicit reasoning, and actively update prior beliefs based on the verification evidence. EVU is designed as a unified intervention mechanism that generates textual belief states explicitly, and can be integrated into both prompting-based and training-based agent reasoning methods. Extensive experiments across three embodied benchmarks demonstrate that EVU consistently yields substantial gains in task success rates. Further analyses validate that our approach effectively mitigates belief inertia, advancing the development of more robust embodied agents. Our code is available at https://github.com/WangHanLinHenry/EVU.