Exploring Conditions for Diffusion models in Robotic Control

arXiv cs.RO / 4/9/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies how to use pre-trained text-to-image diffusion models to produce task-adaptive visual representations for imitation learning in robotics without fine-tuning the diffusion model itself.
  • It finds that directly applying textual conditions that work well in other vision tasks can produce minimal or even negative improvements in robotic control due to a domain gap between diffusion training data and robot environments.
  • The authors argue that effective conditioning must account for the dynamic, fine-grained visual information specific to control, rather than relying on naive text prompts.
  • They propose ORCA, which uses learnable task prompts that adapt to the control environment and visual prompts designed to capture frame-specific details.
  • ORCA achieves state-of-the-art results across multiple robotic control benchmarks, outperforming prior approaches that use frozen pre-trained representations.

Abstract

While pre-trained visual representations have significantly advanced imitation learning, they are often task-agnostic as they remain frozen during policy learning. In this work, we explore leveraging pre-trained text-to-image diffusion models to obtain task-adaptive visual representations for robotic control, without fine-tuning the model itself. However, we find that naively applying textual conditions - a successful strategy in other vision domains - yields minimal or even negative gains in control tasks. We attribute this to the domain gap between the diffusion model's training data and robotic control environments, leading us to argue for conditions that consider the specific, dynamic visual information required for control. To this end, we propose ORCA, which introduces learnable task prompts that adapt to the control environment and visual prompts that capture fine-grained, frame-specific details. Through facilitating task-adaptive representations with our newly devised conditions, our approach achieves state-of-the-art performance on various robotic control benchmarks, significantly surpassing prior methods.