Dynamical Priors as a Training Objective in Reinforcement Learning

arXiv cs.LG / 4/24/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that standard reinforcement learning can achieve high reward while still producing temporally incoherent behaviors like abrupt confidence changes, oscillations, or inactivity.
  • It proposes Dynamical Prior Reinforcement Learning (DP-RL), which adds an auxiliary loss to policy-gradient training based on external state dynamics that encode evidence accumulation and hysteresis.
  • DP-RL is designed to work without changing the reward function, the environment, or the policy architecture, instead shaping how action probabilities evolve over time during training.
  • Experiments on three minimal environments show that the dynamical priors change decision trajectories in task-dependent ways and yield temporally structured behavior beyond what generic smoothing could explain.
  • The authors conclude that the choice of training objectives can directly control the temporal “geometry” of an RL agent’s decision-making process.

Abstract

Standard reinforcement learning (RL) optimizes policies for reward but imposes few constraints on how decisions evolve over time. As a result, policies may achieve high performance while exhibiting temporally incoherent behavior such as abrupt confidence shifts, oscillations, or degenerate inactivity. We introduce Dynamical Prior Reinforcement Learning (DP-RL), a training framework that augments policy gradient learning with an auxiliary loss derived from external state dynamics that implement evidence accumulation and hysteresis. Without modifying the reward, environment, or policy architecture, this prior shapes the temporal evolution of action probabilities during learning. Across three minimal environments, we show that dynamical priors systematically alter decision trajectories in task-dependent ways, promoting temporally structured behavior that cannot be explained by generic smoothing. These results demonstrate that training objectives alone can control the temporal geometry of decision-making in RL agents.