Active Reward Machine Inference From Raw State Trajectories

arXiv cs.RO / 4/10/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper presents a method to learn reward machines directly from raw state trajectories and policy information without needing reward, label, or reward-machine node observations.
  • It argues that, in this information-scarce setting, trajectory data alone can be sufficient to infer the automaton-like reward structure required for multi-stage task specification.
  • The approach is extended to an active learning framework that incrementally queries additional trajectory extensions to improve both data efficiency and computational efficiency.
  • Experiments on grid-world environments demonstrate the feasibility of the learned reward machines under the proposed assumptions.

Abstract

Reward machines are automaton-like structures that capture the memory required to accomplish a multi-stage task. When combined with reinforcement learning or optimal control methods, they can be used to synthesize robot policies to achieve such tasks. However, specifying a reward machine by hand, including a labeling function capturing high-level features that the decisions are based on, can be a daunting task. This paper deals with the problem of learning reward machines directly from raw state and policy information. As opposed to existing works, we assume no access to observations of rewards, labels, or machine nodes, and show what trajectory data is sufficient for learning the reward machine in this information-scarce regime. We then extend the result to an active learning setting where we incrementally query trajectory extensions to improve data (and indirectly computational) efficiency. Results are demonstrated with several grid world examples.