Mitigating Value Hallucination in Dyna Planning via Multistep Predecessor Models

arXiv stat.ML / 4/7/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • Dyna-style reinforcement learning can fail when environment model errors cause simulated states to produce misleading value estimates that harm the learned control policy.
  • The paper proposes the “Hallucinated Value Hypothesis (HVH),” arguing that bootstrapping real-state values toward simulated-state values can lead to incorrect action values and degraded behavior.
  • It surveys a design space of Dyna variants across successor vs. predecessor models (forward vs. backward simulation) and one-step vs. multi-step updates.
  • The authors introduce and evaluate the previously underexplored variant of using predecessor models with multi-step updates, finding it avoids the failure mode suggested by HVH.
  • Experimental results indicate the HVH is supported and that predecessor models with multi-step updates are a promising route to making Dyna more robust to model error.

Abstract

Dyna-style reinforcement learning (RL) agents improve sample efficiency over model-free RL agents by updating the value function with simulated experience generated by an environment model. However, it is often difficult to learn accurate models of environment dynamics, and even small errors may result in failure of Dyna agents. In this paper, we highlight that one potential cause of that failure is bootstrapping off of the values of simulated states, and introduce a new Dyna algorithm to avoid this failure. We discuss a design space of Dyna algorithms, based on using successor or predecessor models -- simulating forwards or backwards -- and using one-step or multi-step updates. Three of the variants have been explored, but surprisingly the fourth variant has not: using predecessor models with multi-step updates. We present the \emph{Hallucinated Value Hypothesis} (HVH): updating the values of real states towards values of simulated states can result in misleading action values which adversely affect the control policy. We discuss and evaluate all four variants of Dyna amongst which three update real states toward simulated states -- so potentially toward hallucinated values -- and our proposed approach, which does not. The experimental results provide evidence for the HVH, and suggest that using predecessor models with multi-step updates is a promising direction toward developing Dyna algorithms that are more robust to model error.