Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

DriveDreamer-2: LLM-Enhanced World Models for Diverse Driving Video Generation

About

World models have demonstrated superiority in autonomous driving, particularly in the generation of multi-view driving videos. However, significant challenges still exist in generating customized driving videos. In this paper, we propose DriveDreamer-2, which builds upon the framework of DriveDreamer and incorporates a Large Language Model (LLM) to generate user-defined driving videos. Specifically, an LLM interface is initially incorporated to convert a user's query into agent trajectories. Subsequently, a HDMap, adhering to traffic regulations, is generated based on the trajectories. Ultimately, we propose the Unified Multi-View Model to enhance temporal and spatial coherence in the generated driving videos. DriveDreamer-2 is the first world model to generate customized driving videos, it can generate uncommon driving videos (e.g., vehicles abruptly cut in) in a user-friendly manner. Besides, experimental results demonstrate that the generated videos enhance the training of driving perception methods (e.g., 3D detection and tracking). Furthermore, video generation quality of DriveDreamer-2 surpasses other state-of-the-art methods, showcasing FID and FVD scores of 11.2 and 55.7, representing relative improvements of 30% and 50%.

Guosheng Zhao, Xiaofeng Wang, Zheng Zhu, Xinze Chen, Guan Huang, Xiaoyi Bao, Xingang Wang• 2024

Related benchmarks

TaskDatasetResultRank
3D Multi-Object TrackingnuScenes (val)
AMOTA31.3
115
Video GenerationnuScenes (val)
FVD55.7
37
Video PredictionnuScenes (val)
FID25
16
Camera GenerationnuScenes v1.0-trainval (val)
FID25
11
Camera GenerationnuScenes (val)
FID11.2
10
Generative World Model Attack EvaluationDriveDreamer v2 (test)
FID18.4
3
Showing 6 of 6 rows

Other info

Follow for update