Learning Sidewalk Autopilot from Multi-Scale Imitation with Corrective Behavior Expansion

arXiv cs.RO / 3/25/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper targets imitation-learning control for sidewalk micromobility, arguing that standard IL on fixed offline datasets suffers from compounding errors and weak robustness in complex urban settings.
  • It proposes “corrective behavior expansion” by augmenting teleoperation data with diverse corrective actions and sensor augmentations so the learned policy can recover from its own mistakes.
  • It introduces a multi-scale imitation learning model that learns both short-horizon interactive behaviors and long-horizon, goal-directed intentions using horizon-based trajectory clustering and hierarchical supervision.
  • Real-world experiments are reported to show significant improvements in robustness and generalization across diverse sidewalk scenarios.

Abstract

Sidewalk micromobility is a promising solution for last-mile transportation, but current learning-based control methods struggle in complex urban environments. Imitation learning (IL) learns policies from human demonstrations, yet its reliance on fixed offline data often leads to compounding errors, limited robustness, and poor generalization. To address these challenges, we propose a framework that advances IL through corrective behavior expansion and multi-scale imitation learning. On the data side, we augment teleoperation datasets with diverse corrective behaviors and sensor augmentations to enable the policy to learn to recover from its own mistakes. On the model side, we introduce a multi-scale IL architecture that captures both short-horizon interactive behaviors and long-horizon goal-directed intentions via horizon-based trajectory clustering and hierarchical supervision. Real-world experiments show that our approach significantly improves robustness and generalization in diverse sidewalk scenarios.