Fourier Features Let Networks Learn High Frequency Functions in Low Dimensional Domains
About
We show that passing input points through a simple Fourier feature mapping enables a multilayer perceptron (MLP) to learn high-frequency functions in low-dimensional problem domains. These results shed light on recent advances in computer vision and graphics that achieve state-of-the-art results by using MLPs to represent complex 3D objects and scenes. Using tools from the neural tangent kernel (NTK) literature, we show that a standard MLP fails to learn high frequencies both in theory and in practice. To overcome this spectral bias, we use a Fourier feature mapping to transform the effective NTK into a stationary kernel with a tunable bandwidth. We suggest an approach for selecting problem-specific Fourier features that greatly improves the performance of MLPs for low-dimensional regression tasks relevant to the computer vision and graphics communities.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Reconstruction | Kodak (test) | PSNR29.71 | 33 | |
| Video Compression | UVG standard (full) | Beauty Quality Score32.58 | 24 | |
| Image Reconstruction | DIV2K (val) | PSNR29.771 | 22 | |
| Video Encoding | UVG-HD | PSNR29.53 | 19 | |
| Implicit Video Representation | UVG-HD full 1920x1080 | PSNR (Beauty)32.58 | 18 | |
| Mesh Reconstruction | Thingi32 | gIoU97.7 | 17 | |
| Surface Reconstruction | Surface Reconstruction Benchmark (SRB) 5 noisy range scans | Dist Error (c) vs GT0.22 | 15 | |
| Image Reconstruction | BSDS500 (test) | PSNR (dB)34.68 | 11 | |
| Image Reconstruction | PCP (test) | PSNR (dB)25.46 | 11 | |
| Implicit Representation | Astronaut (test) | MSE5.1 | 10 |