Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

DeepONet: Learning nonlinear operators for identifying differential equations based on the universal approximation theorem of operators

About

While it is widely known that neural networks are universal approximators of continuous functions, a less known and perhaps more powerful result is that a neural network with a single hidden layer can approximate accurately any nonlinear continuous operator. This universal approximation theorem is suggestive of the potential application of neural networks in learning nonlinear operators from data. However, the theorem guarantees only a small approximation error for a sufficient large network, and does not consider the important optimization and generalization errors. To realize this theorem in practice, we propose deep operator networks (DeepONets) to learn operators accurately and efficiently from a relatively small dataset. A DeepONet consists of two sub-networks, one for encoding the input function at a fixed number of sensors $x_i, i=1,\dots,m$ (branch net), and another for encoding the locations for the output functions (trunk net). We perform systematic simulations for identifying two types of operators, i.e., dynamic systems and partial differential equations, and demonstrate that DeepONet significantly reduces the generalization error compared to the fully-connected networks. We also derive theoretically the dependence of the approximation error in terms of the number of sensors (where the input function is defined) as well as the input function type, and we verify the theorem with computational results. More importantly, we observe high-order error convergence in our computational tests, namely polynomial rates (from half order to fourth order) and even exponential convergence with respect to the training dataset size.

Lu Lu, Pengzhan Jin, George Em Karniadakis• 2019

Related benchmarks

TaskDatasetResultRank
Constitutive modelinghyperelasticity (test)
Relative L2 Error11.998
33
Fluid-solid interaction modelingFluid-solid interaction NS+EW Re=4000 (test)
L2 Loss0.259
24
Fluid dynamics modelingFluid dynamics (NS) Re=400 (test)
L2 Loss0.107
24
Fluid-solid interaction modelingFluid-solid interaction NS+EW Re=400 (test)
L2 Loss0.107
24
Solution ReconstructionInverse Burgers' Equation (test)
Relative MAE7.34
21
Learning PDE Solution Operators2D Shallow Water
Mean L2 Relative Error1.11
20
Inverse Problem Subdomain CompletionInverse Burgers' Equation Subdomain (test)
Relative MAE2.51
15
Inverse PDE solvingHelmholtz full observations
Relative Error0.281
14
Learning PDE Solution Operators1D Diffusion-Reaction
Mean L2 Rel Error61
12
Learning PDE Solution OperatorsAllen-Cahn 1D
Mean L2 Relative Error16.53
12
Showing 10 of 36 rows

Other info

Follow for update