Emergent Neural Automaton Policies: Learning Symbolic Structure from Visuomotor Trajectories

arXiv cs.RO / 3/30/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces ENAP (Emergent Neural Automaton Policy) to improve long-horizon robot learning by combining an interpretable discrete planner with continuous control.
  • ENAP infers a Mealy state machine from visuomotor demonstrations using adaptive clustering plus an L* algorithm extension, capturing latent task modes without needing task-specific labels.
  • The learned discrete transitions then condition a low-level reactive residual network that learns continuous actions via behavior cloning (BC).
  • Experiments on complex manipulation and other long-horizon tasks show ENAP outperforms state-of-the-art end-to-end VLA policies by up to 27% in low-data settings and provides an explicit structured representation of robotic intent.

Abstract

Scaling robot learning to long-horizon tasks remains a formidable challenge. While end-to-end policies often lack the structural priors needed for effective long-term reasoning, traditional neuro-symbolic methods rely heavily on hand-crafted symbolic priors. To address the issue, we introduce ENAP (Emergent Neural Automaton Policy), a framework that allows a bi-level neuro-symbolic policy adaptively emerge from visuomotor demonstrations. Specifically, we first employ adaptive clustering and an extension of the L* algorithm to infer a Mealy state machine from visuomotor data, which serves as an interpretable high-level planner capturing latent task modes. Then, this discrete structure guides a low-level reactive residual network to learn precise continuous control via behavior cloning (BC). By explicitly modeling the task structure with discrete transitions and continuous residuals, ENAP achieves high sample efficiency and interpretability without requiring task-specific labels. Extensive experiments on complex manipulation and long-horizon tasks demonstrate that ENAP outperforms state-of-the-art (SoTA) end-to-end VLA policies by up to 27% in low-data regimes, while offering a structured representation of robotic intent (Fig. 1).