V2VNet: Vehicle-to-Vehicle Communication for Joint Perception and Prediction
About
In this paper, we explore the use of vehicle-to-vehicle (V2V) communication to improve the perception and motion forecasting performance of self-driving vehicles. By intelligently aggregating the information received from multiple nearby vehicles, we can observe the same scene from different viewpoints. This allows us to see through occlusions and detect actors at long range, where the observations are very sparse or non-existent. We also show that our approach of sending compressed deep feature map activations achieves high accuracy while satisfying communication bandwidth requirements.
Tsun-Hsuan Wang, Sivabalan Manivasagam, Ming Liang, Bin Yang, Wenyuan Zeng, James Tu, Raquel Urtasun• 2020
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| 3D Object Detection | DAIR-V2X | AP@0.5066.4 | 57 | |
| Object Detection | IRV2V | AP@0.5081.1 | 48 | |
| 3D Object Detection | OPV2V | AP@0.5078.07 | 47 | |
| 3D Object Detection | DAIR-V2X 14 (test) | AP@0.5062.6 | 43 | |
| 3D Object Detection | V2XSet | AP@0.5090.6 | 24 | |
| 3D Object Detection | OPV2V 41 (test) | AP@0.596.66 | 21 | |
| 3D Object Detection | V2XSim 21 (test) | AP@0.588.97 | 21 | |
| 3D Object Detection | DAIR-V2X 46 (test) | AP@0.566.63 | 21 | |
| Cooperative Object Detection | OPV2V | AP@0.591.7 | 18 | |
| 3D Object Detection | V2X-Real (test) | Vehicle AP@0.386.7 | 18 |
Showing 10 of 23 rows