Learning Lifted Action Models from Unsupervised Visual Traces
arXiv cs.AI / 4/22/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper proposes learning “lifted action models” for AI planning from only sequences of state images, assuming that actions are not directly observed.
- It introduces a deep learning framework that jointly trains state prediction, action prediction, and the lifted (symbolic/parameterized) action model in one end-to-end setup.
- To avoid prediction collapse and self-reinforcing errors, the authors add a MILP that finds logically consistent states, actions, and action-model parameters close to the network’s raw predictions.
- MILP-derived pseudo-labels are then fed back into training, and experiments across multiple domains show improved convergence toward globally consistent solutions.
- Overall, the work advances unsupervised/weakly supervised learning of action dynamics needed for real-world planning systems.
