PETR: Position Embedding Transformation for Multi-View 3D Object Detection
About
In this paper, we develop position embedding transformation (PETR) for multi-view 3D object detection. PETR encodes the position information of 3D coordinates into image features, producing the 3D position-aware features. Object query can perceive the 3D position-aware features and perform end-to-end object detection. PETR achieves state-of-the-art performance (50.4% NDS and 44.1% mAP) on standard nuScenes dataset and ranks 1st place on the benchmark. It can serve as a simple yet strong baseline for future research. Code is available at \url{https://github.com/megvii-research/PETR}.
Yingfei Liu, Tiancai Wang, Xiangyu Zhang, Jian Sun• 2022
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| 3D Object Detection | nuScenes (val) | NDS49.6 | 941 | |
| 3D Object Detection | nuScenes (test) | mAP44.5 | 829 | |
| 3D Object Detection | NuScenes v1.0 (test) | mAP44.5 | 210 | |
| 3D Object Detection | nuScenes v1.0 (val) | mAP (Overall)40.3 | 190 | |
| 3D Object Detection | Argoverse 2 (val) | mAP17.6 | 62 | |
| 3D Object Detection | Waymo Open Dataset LEVEL_1 (val) | 3D AP20.9 | 46 | |
| Object Detection | nuScenes (val) | mAP37 | 41 | |
| 3D Object Detection | Waymo (val) | -- | 38 | |
| 3D Visual Grounding | NuInteract (test) | Precision55.8 | 16 | |
| 3D Object Detection | nuScenes Rainy (val) | mAP41.9 | 13 |
Showing 10 of 14 rows