Self-calibrating Deep Photometric Stereo Networks
About
This paper proposes an uncalibrated photometric stereo method for non-Lambertian scenes based on deep learning. Unlike previous approaches that heavily rely on assumptions of specific reflectances and light source distributions, our method is able to determine both shape and light directions of a scene with unknown arbitrary reflectances observed under unknown varying light directions. To achieve this goal, we propose a two-stage deep learning architecture, called SDPS-Net, which can effectively take advantage of intermediate supervision, resulting in reduced learning difficulty compared to a single-stage model. Experiments on both synthetic and real datasets show that our proposed approach significantly outperforms previous uncalibrated photometric stereo methods.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Photometric Stereo | DiLiGenT (All 96 images) | Ball Error2.8 | 20 | |
| Surface Normal Estimation | DiLiGenT 1.0 (full) | BALL Error2.77 | 10 | |
| Light Calibration | DiLiGenT standard (test) | BALL Directional Error3.27 | 7 | |
| Surface Normal Estimation | DiLiGenT10^2 BALL 1.0 (test) | POM Angular Error4.4 | 5 | |
| Geometry Reconstruction | Our Synthetic RGB Dataset 1.0 (test) | Depth L10.9163 | 5 | |
| Surface Normal Estimation | DiLiGenT10^2 BUNNY 1.0 (test) | POM Error29 | 5 |