Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

MVSAnywhere: Zero-Shot Multi-View Stereo

About

Computing accurate depth from multiple views is a fundamental and longstanding challenge in computer vision. However, most existing approaches do not generalize well across different domains and scene types (e.g. indoor vs. outdoor). Training a general-purpose multi-view stereo model is challenging and raises several questions, e.g. how to best make use of transformer-based architectures, how to incorporate additional metadata when there is a variable number of input views, and how to estimate the range of valid depths which can vary considerably across different scenes and is typically not known a priori? To address these issues, we introduce MVSA, a novel and versatile Multi-View Stereo architecture that aims to work Anywhere by generalizing across diverse domains and depth ranges. MVSA combines monocular and multi-view cues with an adaptive cost volume to deal with scale-related issues. We demonstrate state-of-the-art zero-shot depth estimation on the Robust Multi-View Depth Benchmark, surpassing existing multi-view stereo and monocular baselines.

Sergio Izquierdo, Mohamed Sayed, Michael Firman, Guillermo Garcia-Hernando, Daniyar Turmukhambetov, Javier Civera, Oisin Mac Aodha, Gabriel Brostow, Jamie Watson• 2025

Related benchmarks

TaskDatasetResultRank
Depth EstimationDDAD (val)--
31
Depth EstimationKITTI (val)
MAE1.78
11
Depth EstimationWaymo (val)
MAE3.3
7
Video Depth EstimationKITTI
AbsRel10.37
3
Showing 4 of 4 rows

Other info

Follow for update