Knowledge-Guided Manipulation Using Multi-Task Reinforcement Learning

arXiv cs.RO / 3/26/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes KG-M3PO, a knowledge-graph-based, multi-task model-based policy optimization framework for robotic manipulation under partial observability.
  • It augments egocentric vision with an online 3D scene graph that grounds open-vocabulary detections into a metric, relational representation updated dynamically at every step.
  • A graph neural encoder is trained end-to-end with the reinforcement learning objective, so relational features are shaped directly by manipulation control performance.
  • The approach fuses multiple modalities—visual, proprioceptive, linguistic, and graph-based—into a shared latent space and uses lightweight graph queries for compact, semantically informed policy conditioning.
  • Experiments on manipulation benchmarks with occlusions, distractors, and layout shifts show improved success rates, better sample efficiency, and stronger generalization versus strong baselines.

Abstract

This paper introduces Knowledge Graph based Massively Multi-task Model-based Policy Optimization (KG-M3PO), a framework for multi-task robotic manipulation in partially observable settings that unifies Perception, Knowledge, and Policy. The method augments egocentric vision with an online 3D scene graph that grounds open-vocabulary detections into a metric, relational representation. A dynamic-relation mechanism updates spatial, containment, and affordance edges at every step, and a graph neural encoder is trained end-to-end through the RL objective so that relational features are shaped directly by control performance. Multiple observation modalities (visual, proprioceptive, linguistic, and graph-based) are encoded into a shared latent space, upon which the RL agent operates to drive the control loop. The policy conditions on lightweight graph queries alongside visual and proprioceptive inputs, yielding a compact, semantically informed state for decision making. Experiments on a suite of manipulation tasks with occlusions, distractors, and layout shifts demonstrate consistent gains over strong baselines: the knowledge-conditioned agent achieves higher success rates, improved sample efficiency, and stronger generalization to novel objects and unseen scene configurations. These results support the premise that structured, continuously maintained world knowledge is a powerful inductive bias for scalable, generalizable manipulation: when the knowledge module participates in the RL computation graph, relational representations align with control, enabling robust long-horizon behavior under partial observability.