Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

$\pi^3$: Permutation-Equivariant Visual Geometry Learning

About

We introduce $\pi^3$, a feed-forward neural network that offers a novel approach to visual geometry reconstruction, breaking the reliance on a conventional fixed reference view. Previous methods often anchor their reconstructions to a designated viewpoint, an inductive bias that can lead to instability and failures if the reference is suboptimal. In contrast, $\pi^3$ employs a fully permutation-equivariant architecture to predict affine-invariant camera poses and scale-invariant local point maps without any reference frames. This design not only makes our model inherently robust to input ordering, but also leads to higher accuracy and performance. These advantages enable our simple and bias-free approach to achieve state-of-the-art performance on a wide range of tasks, including camera pose estimation, monocular/video depth estimation, and dense point map reconstruction. Code and models are publicly available.

Yifan Wang, Jianjun Zhou, Haoyi Zhu, Wenzheng Chang, Yang Zhou, Zizun Li, Junyi Chen, Jiangmiao Pang, Chunhua Shen, Tong He• 2025

Related benchmarks

TaskDatasetResultRank
Monocular Depth EstimationKITTI
Abs Rel0.06
161
Monocular Depth EstimationNYU V2--
113
Video Depth EstimationSintel
Relative Error (Rel)0.21
109
Video Depth EstimationBONN
Relative Error (Rel)0.043
103
Camera pose estimationSintel
ATE0.073
92
Camera pose estimationScanNet
ATE RMSE (Avg.)0.03
61
Camera pose estimationTUM dynamics
RRE0.309
57
Video Depth EstimationSintel (test)
Delta 1 Accuracy68.4
57
Video Depth EstimationKITTI
Abs Rel0.037
47
3D ReconstructionNeural RGB-D (NRGBD)
Acc Mean0.026
38
Showing 10 of 67 rows

Other info

Follow for update