Agent-Centric Visual Reinforcement Learning under Dynamic Perturbations

arXiv cs.RO / 4/28/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies how visual reinforcement learning (RL) policies degrade under dynamic, non-stationary visual perturbations such as unpredictable corruption-type shifts.
  • It introduces the Visual Degraded Control Suite (VDCS), extending DeepMind Control Suite with Markov-switching degradations to benchmark robustness in realistic changing conditions.
  • Experiments show that existing methods suffer severe performance drops, and the authors prove theoretically (via information-theoretic analysis) that reconstruction-based objectives cause perturbation artifacts to leak into latent representations.
  • To address this, the paper proposes ACO-MoE (Agent-Centric Observations with Mixture-of-Experts), which uses specialized agent-centric restoration experts to decouple perturbation recovery from task-relevant perception.
  • On VDCS and related generalization tests, ACO-MoE substantially improves robustness, recovering 95.3% of clean performance under Markov-switching corruptions and achieving state-of-the-art results on DMControl Generalization benchmarks.

Abstract

Visual reinforcement learning aims to empower an agent to learn policies from visual observations, yet it remains vulnerable to dynamic visual perturbations, such as unpredictable shifts in corruption types. To systematically study this, we introduce the Visual Degraded Control Suite (VDCS), a benchmark extending DeepMind Control Suite with Markov-switching degradations to simulate non-stationary real-world perturbations. Experiments on VDCS reveal severe performance degradation in existing methods. We theoretically prove via information-theoretic analysis that this failure stems from reconstruction-based objectives inevitably entangling perturbation artifacts into latent representations. To mitigate this negative impact, we propose Agent-Centric Observations with Mixture-of-Experts (ACO-MoE) to robustify visual RL against perturbations. The proposed framework leverages unique agent-centric restoration experts, achieving restoration from corruptions and task-relevant foreground extraction, thereby decoupling perception from perturbation before being processed by the RL agent. Extensive experiments on VDCS show our ACO-MoE outperforms strong baselines, recovering 95.3% of clean performance under challenging Markov-switching corruptions. Moreover, it achieves SOTA results on DMControl Generalization with random-color and video-background perturbations, demonstrating a high level of robustness.