Refinement of Accelerated Demonstrations via Incremental Iterative Reference Learning Control for Fast Contact-Rich Imitation Learning

arXiv cs.RO / 4/21/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper tackles how to generate fast demonstrations for contact-rich manipulation imitation learning, noting that naive time acceleration distorts contact dynamics and increases tracking errors.
  • It introduces Incremental Iterative Reference Learning Control (I2RLC), which adapts the reference trajectory using IRLC while gradually increasing execution speed to improve stability and fidelity.
  • Experiments on real robots (whiteboard erasing and peg-in-hole) show that both IRLC and I2RLC can produce up to 10x faster demonstrations with lower tracking error, while I2RLC achieves about 22.5% better spatial similarity to the original trajectories versus IRLC.
  • Using the refined trajectories to train imitation learning policies yields faster execution and 100% success on peg-in-hole for both seen and unseen positions, with I2RLC-trained policies generating lower contact forces than IRLC-trained ones.
  • Overall, the results suggest that combining incremental speed scheduling with reference adaptation is an effective approach for practical fast contact-rich imitation learning.

Abstract

Fast execution of contact-rich manipulation is critical for practical deployment, yet providing fast demonstrations for imitation learning (IL) remains challenging: humans cannot demonstrate at high speed, and naively accelerating demonstrations alters contact dynamics and induces large tracking errors. We present a method to autonomously refine time-accelerated demonstrations by repurposing Iterative Reference Learning Control (IRLC) to iteratively update the reference trajectory from observed tracking errors. However, applying IRLC directly at high speed tends to produce larger early-iteration errors and less stable transients. To address this issue, we propose Incremental Iterative Reference Learning Control (I2RLC), which gradually increases the speed while updating the reference, yielding high-fidelity trajectories. We validate on real-robot whiteboard erasing and peg-in-hole tasks using a teleoperation setup with a compliance-controlled follower and a 3D-printed haptic leader. Both IRLC and I2RLC achieve up to 10x faster demonstrations with reduced tracking error; moreover, I2RLC improves spatial similarity to the original trajectories by 22.5% on average over IRLC across three tasks and multiple speeds (3x-10x). We then use the refined trajectories to train IL policies; the resulting policies execute faster than the demonstrations and achieve 100% success rates in the peg-in-hole task at both seen and unseen positions, with I2RLC-trained policies exhibiting lower contact forces than those trained on IRLC-refined demonstrations. These results indicate that gradual speed scheduling coupled with reference adaptation provides a practical path to fast, contact-rich IL.