Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Improved Operator Learning by Orthogonal Attention

About

Neural operators, as an efficient surrogate model for learning the solutions of PDEs, have received extensive attention in the field of scientific machine learning. Among them, attention-based neural operators have become one of the mainstreams in related research. However, existing approaches overfit the limited training data due to the considerable number of parameters in the attention mechanism. To address this, we develop an orthogonal attention based on the eigendecomposition of the kernel integral operator and the neural approximation of eigenfunctions. The orthogonalization naturally poses a proper regularization effect on the resulting neural operator, which aids in resisting overfitting and boosting generalization. Experiments on six standard neural operator benchmark datasets comprising both regular and irregular geometries show that our method can outperform competing baselines with decent margins.

Zipeng Xiao, Zhongkai Hao, Bokai Lin, Zhijie Deng, Hang Su• 2023

Related benchmarks

TaskDatasetResultRank
PDE solvingDarcy-Flow 2d (test)
Relative MSE0.0094
33
PDE solvingNavier-Stokes Regular Grid (test)
Relative L2 Error0.1195
25
PDE solvingDarcy Regular Grid (test)
Relative L2 Error0.0076
25
PDE solvingAirfoil Structured Mesh (test)
Relative L2 Error0.0061
23
PDE solvingPipe Structured Mesh (test)
Relative L2 Error0.0052
23
Forward PDE solvingAirfoil
Relative L20.61
21
Forward PDE solvingPlasticity
Relative L2 Error0.0048
21
Forward PDE solvingPipe
Relative L2 Error0.0052
20
Forward PDE solvingElasticity
Relative L2 Error0.0118
19
PDE solvingNavier-Stokes 2D (test)
Relative MSE Loss0.793
18
Showing 10 of 25 rows

Other info

Follow for update