Domain-invariant Stereo Matching Networks
About
State-of-the-art stereo matching networks have difficulties in generalizing to new unseen environments due to significant domain differences, such as color, illumination, contrast, and texture. In this paper, we aim at designing a domain-invariant stereo matching network (DSMNet) that generalizes well to unseen scenes. To achieve this goal, we propose i) a novel "domain normalization" approach that regularizes the distribution of learned representations to allow them to be invariant to domain differences, and ii) a trainable non-local graph-based filter for extracting robust structural and geometric representations that can further enhance domain-invariant generalizations. When trained on synthetic data and generalized to real test sets, our model performs significantly better than all state-of-the-art models. It even outperforms some deep learning models (e.g. MC-CNN) fine-tuned with test-domain data.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Stereo Matching | KITTI 2015 (test) | -- | 144 | |
| Stereo Matching | KITTI 2015 | -- | 118 | |
| Stereo Matching | KITTI 2012 | -- | 81 | |
| Stereo Matching | KITTI 2012 (test) | -- | 76 | |
| Stereo Matching | ETH3D | bad 1.00.062 | 51 | |
| Stereo Matching | Middlebury (test) | 3PE13.8 | 47 | |
| Stereo Matching | Middlebury | Bad Pixel Rate (Thresh 2.0)8.1 | 34 | |
| Stereo Matching | ETH3D (test) | Error Rate (Th=1.0)6.2 | 30 | |
| Stereo Matching | KITTI 15 | D1 Error (%)5.5 | 27 | |
| Stereo Matching | ETH3D (train) | Bad 1.0 Rate6.2 | 23 |