Fourier Features Let Networks Learn High Frequency Functions in Low Dimensional Domains
About
We show that passing input points through a simple Fourier feature mapping enables a multilayer perceptron (MLP) to learn high-frequency functions in low-dimensional problem domains. These results shed light on recent advances in computer vision and graphics that achieve state-of-the-art results by using MLPs to represent complex 3D objects and scenes. Using tools from the neural tangent kernel (NTK) literature, we show that a standard MLP fails to learn high frequencies both in theory and in practice. To overcome this spectral bias, we use a Fourier feature mapping to transform the effective NTK into a stationary kernel with a tunable bandwidth. We suggest an approach for selecting problem-specific Fourier features that greatly improves the performance of MLPs for low-dimensional regression tasks relevant to the computer vision and graphics communities.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Video Compression | UVG standard (full) | Beauty Quality Score32.58 | 24 | |
| Image Reconstruction | DIV2K (val) | PSNR29.771 | 22 | |
| Video Encoding | UVG-HD | PSNR29.53 | 19 | |
| Implicit Video Representation | UVG-HD full 1920x1080 | PSNR (Beauty)32.58 | 18 | |
| Mesh Reconstruction | Thingi32 | gIoU97.7 | 17 | |
| Surface Reconstruction | Surface Reconstruction Benchmark (SRB) 5 noisy range scans | Dist Error (c) vs GT0.22 | 15 | |
| Implicit Representation | Astronaut (test) | MSE5.1 | 10 | |
| Reconstruction | CaCO3 16-view | PSNR29.01 | 10 | |
| Reconstruction | CaCO3 (14-view) | PSNR27.1 | 10 | |
| Reconstruction | CaCO3 12-view | PSNR25.08 | 10 |