Relighting4D: Neural Relightable Human from Videos
About
Human relighting is a highly desirable yet challenging task. Existing works either require expensive one-light-at-a-time (OLAT) captured data using light stage or cannot freely change the viewpoints of the rendered body. In this work, we propose a principled framework, Relighting4D, that enables free-viewpoints relighting from only human videos under unknown illuminations. Our key insight is that the space-time varying geometry and reflectance of the human body can be decomposed as a set of neural fields of normal, occlusion, diffuse, and specular maps. These neural fields are further integrated into reflectance-aware physically based rendering, where each vertex in the neural field absorbs and reflects the light from the environment. The whole framework can be learned from videos in a self-supervised manner, with physically informed priors designed for regularization. Extensive experiments on both real and synthetic datasets demonstrate that our framework is capable of relighting dynamic human actors with free-viewpoints.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Diffuse Albedo Estimation | SyntheticHuman++ poses (train) | PSNR25.37 | 10 | |
| Relighting | SyntheticHuman++ poses (train) | PSNR22.13 | 10 | |
| Normal estimation | SyntheticHuman++ poses (train) | Angular Error (Degree)26.17 | 10 | |
| Visibility Estimation | SyntheticHuman++ (train poses) | PSNR17.1 | 8 | |
| Relighting | Synthetic Dataset novel pose | PSNR21.9 | 2 |