On universal approximation and error bounds for Fourier Neural Operators
About
Fourier neural operators (FNOs) have recently been proposed as an effective framework for learning operators that map between infinite-dimensional spaces. We prove that FNOs are universal, in the sense that they can approximate any continuous operator to desired accuracy. Moreover, we suggest a mechanism by which FNOs can approximate operators associated with PDEs efficiently. Explicit error bounds are derived to show that the size of the FNO, approximating operators associated with a Darcy type elliptic PDE and with the incompressible Navier-Stokes equations of fluid dynamics, only increases sub (log)-linearly in terms of the reciprocal of the error. Thus, FNOs are shown to efficiently approximate operators arising in a large class of PDEs.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| PDE solving | Darcy flow noisy observations, sigma_max^2 = 0.001 | Test Error0.0125 | 11 | |
| PDE solving | Darcy flow clean data | Test Error0.008 | 11 | |
| PDE solving | Darcy flow noisy inputs, sigma_max^2 = 0.001 | Test Error0.0079 | 11 | |
| Steady-state PDE Solving | Navier-Stokes viscosity=0.001 Noisy observations (sigma_max^o)^2 = 0.001 (test) | Test Error0.152 | 8 | |
| Solving Darcy Flow | Darcy flow noisy observations, sigma_y_max^2 = 0.001 (test) | Test Error0.0125 | 8 | |
| Steady-state PDE Solving | Navier-Stokes viscosity=0.001 Clean sigma_max^2 = 0 (test) | Test Error15.7 | 8 | |
| Steady-state PDE Solving | Navier-Stokes Noisy inputs (viscosity=0.001) (sigma_max^i)^2 = 0.001 (test) | Test Error0.176 | 8 | |
| Solving Darcy Flow | Darcy flow clean (test) | Test Error0.008 | 8 | |
| Solving Darcy Flow | Darcy flow noisy inputs, sigma_x_max^2 = 0.001 (test) | Test Error0.0079 | 8 |