Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Enhancing End-to-End Autonomous Driving with Latent World Model

About

In autonomous driving, end-to-end planners directly utilize raw sensor data, enabling them to extract richer scene features and reduce information loss compared to traditional planners. This raises a crucial research question: how can we develop better scene feature representations to fully leverage sensor data in end-to-end driving? Self-supervised learning methods show great success in learning rich feature representations in NLP and computer vision. Inspired by this, we propose a novel self-supervised learning approach using the LAtent World model (LAW) for end-to-end driving. LAW predicts future scene features based on current features and ego trajectories. This self-supervised task can be seamlessly integrated into perception-free and perception-based frameworks, improving scene feature learning and optimizing trajectory prediction. LAW achieves state-of-the-art performance across multiple benchmarks, including real-world open-loop benchmark nuScenes, NAVSIM, and simulator-based closed-loop benchmark CARLA. The code is released at https://github.com/BraveGroup/LAW.

Yingyan Li, Lue Fan, Jiawei He, Yuqi Wang, Yuntao Chen, Zhaoxiang Zhang, Tieniu Tan• 2024

Related benchmarks

TaskDatasetResultRank
Open-loop planningnuScenes
L2 Error (Avg)0.61
103
Autonomous Driving PlanningNAVSIM v1
NC96.4
86
PlanningnuScenes (val)
Collision Rate (Avg)19
80
Open-loop planningnuScenes v1.0 (val)
L2 (1s)0.31
71
Autonomous Driving PlanningNAVSIM (navtest)
NC97.2
68
Autonomous Driving PlanningNAVSIM v1 (test)
NC96.4
59
Autonomous DrivingNAVSIM (test)
PDMS84.6
48
Autonomous DrivingCARLA Town05 (Long)
DS70.1
46
PlanningNAVSIM (test)
PDMS84.6
44
PlanningnuScenes v1.0-trainval (val)--
39
Showing 10 of 25 rows

Other info

Code

Follow for update