Cascading Convolutional Color Constancy
About
Regressing the illumination of a scene from the representations of object appearances is popularly adopted in computational color constancy. However, it's still challenging due to intrinsic appearance and label ambiguities caused by unknown illuminants, diverse reflection property of materials and extrinsic imaging factors (such as different camera sensors). In this paper, we introduce a novel algorithm by Cascading Convolutional Color Constancy (in short, C4) to improve robustness of regression learning and achieve stable generalization capability across datasets (different cameras and scenes) in a unique framework. The proposed C4 method ensembles a series of dependent illumination hypotheses from each cascade stage via introducing a weighted multiply-accumulate loss function, which can inherently capture different modes of illuminations and explicitly enforce coarse-to-fine network optimization. Experimental results on the public Color Checker and NUS 8-Camera benchmarks demonstrate superior performance of the proposed algorithm in comparison with the state-of-the-art methods, especially for more difficult scenes.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Color Constancy | NUS-8 (three-fold cross val) | Mean Angular Error (MAE)1.96 | 31 | |
| Color Constancy | NCC (in-dataset) | Median Error6.24 | 29 | |
| Color Constancy | Gehler (3-fold cross validation) | Mean Error1.35 | 26 | |
| Color Constancy | LEVI (in-dataset) | Median Error7.01 | 24 | |
| Color Constancy | NCC trained on LEVI (test) | Median Error13.8 | 10 | |
| Color Constancy | Gehler-Shi | Median Error5.62 | 7 | |
| Auto White Balance | NCC → LEVI | Median Error15.09 | 5 |