Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

MoE3D: A Mixture-of-Experts Module for 3D Reconstruction

About

We propose a simple yet effective approach to enhance the performance of feed-forward 3D reconstruction models. Existing methods often struggle near depth discontinuities, where standard regression losses encourage spatial averaging and thus blur sharp boundaries. To address this issue, we introduce a mixture-of-experts formulation that handles uncertainty at depth boundaries by combining multiple smooth depth predictions. A softmax weighting head dynamically selects among these hypotheses on a per-pixel basis. By integrating our mixture model into a pre-trained state-of-the-art 3D model, we achieve a substantial reduction of boundary artifacts and gains in overall reconstruction accuracy. Notably, our approach is highly compute efficient, delivering generalizable improvements even when fine-tuned on a small subset of training data while incurring only negligible additional inference computation, suggesting a promising direction for lightweight and accurate 3D reconstruction.

Zichen Wang, Ang Cao, Liam J. Wang, Jeong Joon Park• 2026

Related benchmarks

TaskDatasetResultRank
3D Reconstruction7 Scenes
Accuracy Median1.5
94
3D ReconstructionNeural RGB-D (NRGBD)
Acc Mean0.055
88
Geometric boundary sharpnessNYU V2
mIoU19.4
4
Geometric boundary sharpnessSintel
mIoU19.4
4
Geometric boundary sharpnessNRGBD
mIoU40.2
4
Showing 5 of 5 rows

Other info

Follow for update