Dynamic Lookahead Distance via Reinforcement Learning-Based Pure Pursuit for Autonomous Racing

arXiv cs.RO / 3/31/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses a core limitation of Pure Pursuit path tracking: performance depends heavily on the chosen fixed lookahead distance, which trades off cornering stability and curve accuracy.
  • It proposes a hybrid controller where a PPO reinforcement learning agent dynamically outputs the lookahead distance using vehicle speed and multi-horizon curvature features while the classical Pure Pursuit controller handles tracking.
  • Training uses Stable-Baselines3 in the F1TENTH Gym simulator with PPO stability measures (KL penalty and learning-rate decay), and deployment is done in a ROS2-based system.
  • Simulation and preliminary real-car tests show improved lap times and more reliable repeated laps on unseen tracks, with observed behavior of increasing lookahead on straights and decreasing it in curves.
  • The approach demonstrates sim-to-real transfer on a 1:10-scale autonomous racing platform, suggesting the learned lookahead adaptation can generalize to new tracks and hardware without retraining (“zero-shot” transfer).

Abstract

Pure Pursuit (PP) is a widely used path-tracking algorithm in autonomous vehicles due to its simplicity and real-time performance. However, its effectiveness is sensitive to the choice of lookahead distance: shorter values improve cornering but can cause instability on straights, while longer values improve smoothness but reduce accuracy in curves. We propose a hybrid control framework that integrates Proximal Policy Optimization (PPO) with the classical Pure Pursuit controller to adjust the lookahead distance dynamically during racing. The PPO agent maps vehicle speed and multi-horizon curvature features to an online lookahead command. It is trained using Stable-Baselines3 in the F1TENTH Gym simulator with a KL penalty and learning-rate decay for stability, then deployed in a ROS2 environment to guide the controller. Experiments in simulation compare the proposed method against both fixed-lookahead Pure Pursuit and an adaptive Pure Pursuit baseline. Additional real-car experiments compare the learned controller against a fixed-lookahead Pure Pursuit controller. Results show that the learned policy improves lap-time performance and repeated lap completion on unseen tracks, while also transferring zero-shot to hardware. The learned controller adapts the lookahead by increasing it on straights and reducing it in curves, demonstrating effectiveness in augmenting a classical controller by online adaptation of a single interpretable parameter. On unseen tracks, the proposed method achieved 33.16 s on Montreal and 46.05 s on Yas Marina, while tolerating more aggressive speed-profile scaling than the baselines and achieving the best lap times among the tested settings. Initial real-car experiments further support sim-to-real transfer on a 1:10-scale autonomous racing platform