Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Dual Cone Gradient Descent for Training Physics-Informed Neural Networks

About

Physics-informed neural networks (PINNs) have emerged as a prominent approach for solving partial differential equations (PDEs) by minimizing a combined loss function that incorporates both boundary loss and PDE residual loss. Despite their remarkable empirical performance in various scientific computing tasks, PINNs often fail to generate reasonable solutions, and such pathological behaviors remain difficult to explain and resolve. In this paper, we identify that PINNs can be adversely trained when gradients of each loss function exhibit a significant imbalance in their magnitudes and present a negative inner product value. To address these issues, we propose a novel optimization framework, Dual Cone Gradient Descent (DCGD), which adjusts the direction of the updated gradient to ensure it falls within a dual cone region. This region is defined as a set of vectors where the inner products with both the gradients of the PDE residual loss and the boundary loss are non-negative. Theoretically, we analyze the convergence properties of DCGD algorithms in a non-convex setting. On a variety of benchmark equations, we demonstrate that DCGD outperforms other optimization algorithms in terms of various evaluation metrics. In particular, DCGD achieves superior predictive accuracy and enhances the stability of training for failure modes of PINNs and complex PDEs, compared to existing optimally tuned models. Moreover, DCGD can be further improved by combining it with popular strategies for PINNs, including learning rate annealing and the Neural Tangent Kernel (NTK).

Youngsik Hwang, Dong-Young Lim• 2024

Related benchmarks

TaskDatasetResultRank
PDE solvingHelmholtz equation
Relative L2 Error0.23
32
PDE solvingKlein-Gordon equation
Relative L2 Error0.005
15
PDE solvingViscous Burgers' equation
Relative L2 Error0.0104
11
Solving PDEsHelmholtz Equation (test)
Max Relative L2 Error0.0036
11
Solving PDEsBurgers Equation (test)
Max Relative L2 Error0.0016
11
Solving PDEsKlein-Gordon Equation (test)
Max Relative L2 Error0.0068
11
PDE solving5D-Heat equation
Relative L2 Error8.00e-4
10
Integro-Differential Equation SolvingVolterra IDEs
Relative L2 Error0.0011
2
Partial Differential Equation SolvingChaotic Kuramoto-Sivashinsky (KS) equation
Relative L2 Error0.0376
2
Partial Differential Equation SolvingConvection equation
Relative L2 Error0.246
2
Showing 10 of 10 rows

Other info

Code

Follow for update