Incremental Residual Reinforcement Learning Toward Real-World Learning for Social Navigation

arXiv cs.RO / 4/10/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses social navigation for mobile robots, noting that wide regional variation makes simulation-heavy training insufficient for covering real-world pedestrian dynamics and conventions.
  • It proposes Incremental Residual Reinforcement Learning (IRRL), combining lightweight incremental learning (no replay buffer or batch updates) with residual RL that trains only the residual policy relative to a base policy.
  • Simulation results show IRRL can match the performance of standard replay-buffer-based RL methods while outperforming prior incremental learning approaches.
  • Real-world experiments further indicate that robots using IRRL can adapt effectively to previously unseen environments through direct on-robot learning, supporting the method’s practicality under edge-device constraints.

Abstract

As the demand for mobile robots continues to increase, social navigation has emerged as a critical task, driving active research into deep reinforcement learning (RL) approaches. However, because pedestrian dynamics and social conventions vary widely across different regions, simulations cannot easily encompass all possible real-world scenarios. Real-world RL, in which agents learn while operating directly in physical environments, presents a promising solution to this issue. Nevertheless, this approach faces significant challenges, particularly regarding constrained computational resources on edge devices and learning efficiency. In this study, we propose incremental residual RL (IRRL). This method integrates incremental learning, which is a lightweight process that operates without a replay buffer or batch updates, with residual RL, which enhances learning efficiency by training only on the residuals relative to a base policy. Through the simulation experiments, we demonstrated that, despite lacking a replay buffer, IRRL achieved performance comparable to those of conventional replay buffer-based methods and outperformed existing incremental learning approaches. Furthermore, the real-world experiments confirmed that IRRL can enable robots to effectively adapt to previously unseen environments through the real-world learning.