Driving on Registers
About
We present DrivoR, a simple and efficient transformer-based architecture for end-to-end autonomous driving. Our approach builds on pretrained Vision Transformers (ViTs) and introduces camera-aware register tokens that compress multi-camera features into a compact scene representation, significantly reducing downstream computation without sacrificing accuracy. These tokens drive two lightweight transformer decoders that generate and then score candidate trajectories. The scoring decoder learns to mimic an oracle and predicts interpretable sub-scores representing aspects such as safety, comfort, and efficiency, enabling behavior-conditioned driving at inference. Despite its minimal design, DrivoR outperforms or matches strong contemporary baselines across NAVSIM-v1, NAVSIM-v2, and the photorealistic closed-loop HUGSIM benchmark. Our results show that a pure-transformer architecture, combined with targeted token compression, is sufficient for accurate, efficient, and adaptive end-to-end driving. Code and checkpoints will be made available via the project page.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Autonomous Driving | NAVSIM v1 (test) | NC99.1 | 99 | |
| Autonomous Driving Trajectory Planning | NAVSIM navhard-two-stage v2 (test) | Stage 1 NC99.1 | 23 | |
| End-to-end Autonomous Driving | NAVSIM navhard-two-stage v2 (test) | Stage 1 Navigation Completion99.1 | 9 | |
| Closed-loop Autonomous Driving | HUGSIM zero-shot | RC (Easy)80.9 | 4 |