Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Align3R: Aligned Monocular Depth Estimation for Dynamic Videos

About

Recent developments in monocular depth estimation methods enable high-quality depth estimation of single-view images but fail to estimate consistent video depth across different frames. Recent works address this problem by applying a video diffusion model to generate video depth conditioned on the input video, which is training-expensive and can only produce scale-invariant depth values without camera poses. In this paper, we propose a novel video-depth estimation method called Align3R to estimate temporal consistent depth maps for a dynamic video. Our key idea is to utilize the recent DUSt3R model to align estimated monocular depth maps of different timesteps. First, we fine-tune the DUSt3R model with additional estimated monocular depth as inputs for the dynamic scenes. Then, we apply optimization to reconstruct both depth maps and camera poses. Extensive experiments demonstrate that Align3R estimates consistent video depth and camera poses for a monocular video with superior performance than baseline methods.

Jiahao Lu, Tianyu Huang, Peng Li, Zhiyang Dou, Cheng Lin, Zhiming Cui, Zhen Dong, Sai-Kit Yeung, Wenping Wang, Yuan Liu• 2024

Related benchmarks

TaskDatasetResultRank
Video Depth EstimationSintel
Relative Error (Rel)0.263
109
Video Depth EstimationBONN
Relative Error (Rel)0.058
103
Camera pose estimationSintel
ATE0.128
92
Camera pose estimationTUM dynamics
RRE0.321
57
Depth PredictionSintel
AbsRel0.253
32
Video Depth EstimationTUM dynamics
Abs Rel0.104
27
Pose EstimationBONN
ATE0.023
10
Video Depth EstimationPointOdyssey (val)
Abs Rel0.077
8
Video Depth EstimationBonn 5 scenes
Abs Rel0.068
8
Video Depth EstimationFlyingThings3D (test)
Abs Rel0.102
7
Showing 10 of 17 rows

Other info

Code

Follow for update