Egocentric Tactile and Proximity Sensors as Observation Priors for Humanoid Collision Avoidance

arXiv cs.RO / 4/29/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a reinforcement learning framework to learn whole-body, collision-free avoidance for a humanoid robot (H1-2) using egocentric tactile and proximity sensors distributed on its body.
  • It studies how sensor properties—such as coverage, sensor type, and sensing range—affect the avoidance behaviors the robot learns.
  • Using a dodgeball benchmark, the authors ablate (remove/modify) different upper-body sensor configurations to evaluate which sensing assumptions improve performance.
  • The results suggest that raw proximity readings can replace explicit object localization when the sensing range is long enough, and that sparse, non-directional proximity signals can be more sample-efficient than dense, directional ones.
  • Overall, the work provides practical observation “prior” guidance for designing sensor layouts for humanoid collision avoidance.

Abstract

Collision-free motion is often aided by tactile and proximity sensors distributed on the body of the robot due to their resistance to occlusion as opposed to external cameras. However, how to shape the sensor's properties, such as sensing coverage; type; and range, to enable avoidant behavior remains unclear. In this work, we present a reinforcement learning framework for whole-body collision avoidance on a humanoid H1-2 robot and use it to characterize how sensor properties shape learned avoidance behavior. Using dodgeball as a benchmark task, we ablate the properties of sensors distributed across the upper body of the robot and find that raw proximity measurements can substitute for explicit object localization provided the sensing range is sufficient and that sparse non-directional proximity signals outpace dense directional alternatives in sample efficiency.