BXRL: Behavior-Explainable Reinforcement Learning

arXiv cs.LG / 2026/3/26

📰 ニュースSignals & Early TrendsIdeas & Deep AnalysisModels & Research

要点

  • The paper introduces Behavior-Explainable Reinforcement Learning (BXRL), a new RL problem formulation that treats “behavior” as a first-class object rather than an after-the-fact explanation of actions.
  • BXRL formalizes behaviors as measurable functions m: Π → ℝ, letting users specify the action-pattern they care about and quantify how strongly a policy exhibits it.
  • It extends explainable RL querying by enabling “Explain this behavior,” and reframes contrastive questions about preferring action a over a′ into a preference for high behavior measures m(π).
  • The authors analyze three existing explainability methods and outline how they could be adapted to explain behavior, rather than implementing a new XRL technique themselves.
  • They provide a JAX port of the HighwayEnv driving environment to support defining, measuring, and differentiating behaviors with respect to model parameters.

Abstract

A major challenge of Reinforcement Learning is that agents often learn undesired behaviors that seem to defy the reward structure they were given. Explainable Reinforcement Learning (XRL) methods can answer queries such as "explain this specific action", "explain this specific trajectory", and "explain the entire policy". However, XRL lacks a formal definition for behavior as a pattern of actions across many episodes. We provide such a definition, and use it to enable a new query: "Explain this behavior". We present Behavior-Explainable Reinforcement Learning (BXRL), a new problem formulation that treats behaviors as first-class objects. BXRL defines a behavior measure as any function m : \Pi \to \mathbb{R}, allowing users to precisely express the pattern of actions that they find interesting and measure how strongly the policy exhibits it. We define contrastive behaviors that reduce the question "why does the agent prefer a to a'?" to "why is m(\pi) high?" which can be explored with differentiation. We do not implement an explainability method; we instead analyze three existing methods and propose how they could be adapted to explain behavior. We present a port of the HighwayEnv driving environment to JAX, which provides an interface for defining, measuring, and differentiating behaviors with respect to the model parameters.