Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Learning to Tune Pure Pursuit in Autonomous Racing: Joint Lookahead and Steering-Gain Control with PPO

About

Pure Pursuit (PP) is widely used in autonomous racing for real-time path tracking due to its efficiency and geometric clarity, yet performance is highly sensitive to how key parameters-lookahead distance and steering gain-are chosen. Standard velocity-based schedules adjust these only approximately and often fail to transfer across tracks and speed profiles. We propose a reinforcement-learning (RL) approach that jointly chooses the lookahead Ld and a steering gain g online using Proximal Policy Optimization (PPO). The policy observes compact state features (speed and curvature taps) and outputs (Ld, g) at each control step. Trained in F1TENTH Gym and deployed in a ROS 2 stack, the policy drives PP directly (with light smoothing) and requires no per-map retuning. Across simulation and real-car tests, the proposed RL-PP controller that jointly selects (Ld, g) consistently outperforms fixed-lookahead PP, velocity-scheduled adaptive PP, and an RL lookahead-only variant, and it also exceeds a kinematic MPC raceline tracker under our evaluated settings in lap time, path-tracking accuracy, and steering smoothness, demonstrating that policy-guided parameter tuning can reliably improve classical geometry-based control.

Mohamed Elgouhary, Amr S. El-Wakeel• 2026

Related benchmarks

TaskDatasetResultRank
Autonomous RacingYas Marina F1TENTH racetrack (10 consecutive laps)
Mean Lap Time45.1
5
Autonomous RacingMontreal F1TENTH racetrack Zero-shot evaluation
Mean Lap Time32.85
5
Autonomous Racing Lap TimingF1TENTH Real-car real-world v_max = 6 m/s (test)
Mean Lap Time9.46
5
Showing 3 of 3 rows

Other info

Follow for update