Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

TransFuser: Imitation with Transformer-Based Sensor Fusion for Autonomous Driving

About

How should we integrate representations from complementary sensors for autonomous driving? Geometry-based fusion has shown promise for perception (e.g. object detection, motion forecasting). However, in the context of end-to-end driving, we find that imitation learning based on existing sensor fusion methods underperforms in complex driving scenarios with a high density of dynamic agents. Therefore, we propose TransFuser, a mechanism to integrate image and LiDAR representations using self-attention. Our approach uses transformer modules at multiple resolutions to fuse perspective view and bird's eye view feature maps. We experimentally validate its efficacy on a challenging new benchmark with long routes and dense traffic, as well as the official leaderboard of the CARLA urban driving simulator. At the time of submission, TransFuser outperforms all prior work on the CARLA leaderboard in terms of driving score by a large margin. Compared to geometry-based fusion, TransFuser reduces the average collisions per kilometer by 48%.

Kashyap Chitta, Aditya Prakash, Bernhard Jaeger, Zehao Yu, Katrin Renz, Andreas Geiger• 2022

Related benchmarks

TaskDatasetResultRank
Autonomous DrivingNAVSIM v1 (test)
NC97.7
99
PlanningNAVSIM (navtest)
NC99.2
53
Autonomous Driving PlanningNAVSIM (navtest)
NC97.8
50
Autonomous DrivingCARLA Town05 (Long)
DS31
46
Autonomous DrivingLongest6
DS76.91
35
Autonomous DrivingNAVSIM (test)
PDMS83.8
34
Autonomous Driving Trajectory PlanningNAVSIM navhard-two-stage v2 (test)
Stage 1 NC96.2
23
PlanningNAVSIM (test)
PDMS84
22
PlanningNavSim (Navhard)
NC0.962
18
Trajectory PlanningNAVSIM v2 (navhard)
NC Rate96.3
17
Showing 10 of 70 rows

Other info

Code

Follow for update