Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

OpenEMMA: Open-Source Multimodal Model for End-to-End Autonomous Driving

About

Since the advent of Multimodal Large Language Models (MLLMs), they have made a significant impact across a wide range of real-world applications, particularly in Autonomous Driving (AD). Their ability to process complex visual data and reason about intricate driving scenarios has paved the way for a new paradigm in end-to-end AD systems. However, the progress of developing end-to-end models for AD has been slow, as existing fine-tuning methods demand substantial resources, including extensive computational power, large-scale datasets, and significant funding. Drawing inspiration from recent advancements in inference computing, we propose OpenEMMA, an open-source end-to-end framework based on MLLMs. By incorporating the Chain-of-Thought reasoning process, OpenEMMA achieves significant improvements compared to the baseline when leveraging a diverse range of MLLMs. Furthermore, OpenEMMA demonstrates effectiveness, generalizability, and robustness across a variety of challenging driving scenarios, offering a more efficient and effective approach to autonomous driving. We release all the codes in https://github.com/taco-group/OpenEMMA.

Shuo Xing, Chengyuan Qian, Yuping Wang, Hongyuan Hua, Kexin Tian, Yang Zhou, Zhengzhong Tu• 2024

Related benchmarks

TaskDatasetResultRank
Visual Question AnsweringInfoVQA
Accuracy71.4
135
Open-loop planningnuScenes
L2 Error (Avg)2.81
103
Open-loop planningnuScenes v1.0 (val)
L2 (1s)1.45
71
Visual Question AnsweringTallyQA
Accuracy80
49
Trajectory PlanningnuScenes
ST-P3 L2 Error (1s)1.45
49
End-to-end PlanningnuScenes
L2 Error (3s)3.76
34
Open-loop trajectory predictionNuScenes v1.0 (test)
L2 Error (1s)1.49
29
PlanningnuScenes
L2 Error (Avg)2.81
24
End-to-end Motion PlanningnuScenes
L2 Displacement Error (1s)1.45
22
Motion PlanningnuScenes
L2 Error (1s)1.45
15
Showing 10 of 33 rows

Other info

Follow for update