MoE3D: A Mixture-of-Experts Module for 3D Reconstruction
About
We propose a simple yet effective approach to enhance the performance of feed-forward 3D reconstruction models. Existing methods often struggle near depth discontinuities, where standard regression losses encourage spatial averaging and thus blur sharp boundaries. To address this issue, we introduce a mixture-of-experts formulation that handles uncertainty at depth boundaries by combining multiple smooth depth predictions. A softmax weighting head dynamically selects among these hypotheses on a per-pixel basis. By integrating our mixture model into a pre-trained state-of-the-art 3D model, we achieve a substantial reduction of boundary artifacts and gains in overall reconstruction accuracy. Notably, our approach is highly compute efficient, delivering generalizable improvements even when fine-tuned on a small subset of training data while incurring only negligible additional inference computation, suggesting a promising direction for lightweight and accurate 3D reconstruction.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| 3D Reconstruction | Neural RGB-D (NRGBD) | Acc Mean0.055 | 38 | |
| 3D Reconstruction | 7 Scenes | Accuracy Mean3.5 | 32 | |
| Geometric boundary sharpness | NYU V2 | mIoU19.4 | 4 | |
| Geometric boundary sharpness | Sintel | mIoU19.4 | 4 | |
| Geometric boundary sharpness | NRGBD | mIoU40.2 | 4 |