Gradient Alignment in Physics-informed Neural Networks: A Second-Order Optimization Perspective
About
Multi-task learning through composite loss functions is fundamental to modern deep learning, yet optimizing competing objectives remains challenging. We present new theoretical and practical approaches for addressing directional conflicts between loss terms, demonstrating their effectiveness in physics-informed neural networks (PINNs) where such conflicts are particularly challenging to resolve. Through theoretical analysis, we demonstrate how these conflicts limit first-order methods and show that second-order optimization naturally resolves them through implicit gradient alignment. We prove that SOAP, a recently proposed quasi-Newton method, efficiently approximates the Hessian preconditioner, enabling breakthrough performance in PINNs: state-of-the-art results on 10 challenging PDE benchmarks, including the first successful application to turbulent flows with Reynolds numbers up to 10,000, with 2-10x accuracy improvements over existing methods. We also introduce a novel gradient alignment score that generalizes cosine similarity to multiple gradients, providing a practical tool for analyzing optimization dynamics. Our findings establish frameworks for understanding and resolving gradient conflicts, with broad implications for optimization beyond scientific computing.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| PDE solving | Burgers (nu = 1/100pi) (test) | Relative L2 Error4.03e-5 | 6 | |
| Solving nonlinear PDE | Allen-Cahn (test) | Relative L2 Error3.48e-6 | 5 | |
| Solving Time-Dependent PDEs | Burgers | Relative L2 Error1.97e-6 | 4 | |
| Solving Time-Dependent PDEs | Allen-Cahn | Relative L2 Error8.32e-6 | 4 | |
| Solving Time-Dependent PDEs | Korteweg-De Vries | Relative L2 Error4.26e-4 | 4 | |
| Solving Time-Dependent PDEs | Wave | Rel L2 Error2.88e-5 | 4 | |
| Solving Time-Dependent PDEs | Klein-Gordon | Rel L2 Error1.88e-5 | 4 |