Spatiotemporally Consistent Indoor Lighting Estimation with Diffusion Priors
About
Indoor lighting estimation from a single image or video remains a challenge due to its highly ill-posed nature, especially when the lighting condition of the scene varies spatially and temporally. We propose a method that estimates from an input video a continuous light field describing the spatiotemporally varying lighting of the scene. We leverage 2D diffusion priors for optimizing such light field represented as a MLP. To enable zero-shot generalization to in-the-wild scenes, we fine-tune a pre-trained image diffusion model to predict lighting at multiple locations by jointly inpainting multiple chrome balls as light probes. We evaluate our method on indoor lighting estimation from a single image or video and show superior performance over compared baselines. Most importantly, we highlight results on spatiotemporally consistent lighting estimation from in-the-wild videos, which is rarely demonstrated in previous works.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Lighting Estimation | Laval Indoor SV | RMSE (Mirror)0.35 | 4 | |
| Lighting Estimation | Infinigen | RMSE (Mirror)0.34 | 4 | |
| Lighting Estimation | Blender-based Dynamic Scenes Dynamic object (test) | RMSE (Mirr)0.39 | 3 | |
| Lighting Estimation | Blender-based Dynamic Scenes Dynamic camera (test) | RMSE (Mirr)0.39 | 3 | |
| Lighting Estimation | Blender-based Dynamic Scenes Dynamic lighting (test) | RMSE (Mirr)0.39 | 3 | |
| Lighting Estimation | Blender-based Dynamic Scenes Combination (test) | RMSE (Mirr)0.38 | 3 |