Academic

Scaling Laws and Pathologies of Single-Layer PINNs: Network Width and PDE Nonlinearity

arXiv:2603.12556v1 Announce Type: new Abstract: We establish empirical scaling laws for Single-Layer Physics-Informed Neural Networks on canonical nonlinear PDEs. We identify a dual optimization failure: (i) a baseline pathology, where the solution error fails to decrease with network width, even at fixed nonlinearity, falling short of theoretical approximation bounds, and (ii) a compounding pathology, where this failure is exacerbated by nonlinearity. We provide quantitative evidence that a simple separable power law is insufficient, and that the scaling behavior is governed by a more complex, non-separable relationship. This failure is consistent with the concept of spectral bias, where networks struggle to learn the high-frequency solution components that intensify with nonlinearity. We show that optimization, not approximation capacity, is the primary bottleneck, and propose a methodology to empirically measure these complex scaling effects.

F
Faris Chaudhry
· · 1 min read · 20 views

arXiv:2603.12556v1 Announce Type: new Abstract: We establish empirical scaling laws for Single-Layer Physics-Informed Neural Networks on canonical nonlinear PDEs. We identify a dual optimization failure: (i) a baseline pathology, where the solution error fails to decrease with network width, even at fixed nonlinearity, falling short of theoretical approximation bounds, and (ii) a compounding pathology, where this failure is exacerbated by nonlinearity. We provide quantitative evidence that a simple separable power law is insufficient, and that the scaling behavior is governed by a more complex, non-separable relationship. This failure is consistent with the concept of spectral bias, where networks struggle to learn the high-frequency solution components that intensify with nonlinearity. We show that optimization, not approximation capacity, is the primary bottleneck, and propose a methodology to empirically measure these complex scaling effects.

Executive Summary

This article delves into the empirical scaling laws of Single-Layer Physics-Informed Neural Networks (PINNs) on canonical nonlinear Partial Differential Equations (PDEs). The authors identify two pathologies in the optimization process: a baseline pathology where the solution error fails to decrease with network width, and a compounding pathology where this failure is exacerbated by PDE nonlinearity. They provide evidence that a simple power law is insufficient and propose a methodology to measure complex scaling effects. This study offers valuable insights into the limitations of single-layer PINNs and sheds light on the optimization challenges in learning high-frequency solution components. The authors' findings have significant implications for the development of PINNs, particularly in tackling nonlinear PDEs, and highlight the need for more sophisticated network architectures.

Key Points

  • Identification of dual optimization failures in single-layer PINNs
  • Exacerbating effect of PDE nonlinearity on solution error
  • Complexity of scaling behavior beyond separable power laws

Merits

Theoretical significance

The study provides a thorough empirical analysis of single-layer PINNs, shedding light on the optimization challenges and limitations of these networks in tackling nonlinear PDEs.

Methodological innovation

The authors propose a novel methodology to empirically measure complex scaling effects, which could be a valuable contribution to the field of PINNs.

Demerits

Restrictive scope

The study is limited to single-layer PINNs and canonical nonlinear PDEs, which may not be directly applicable to more complex or multi-dimensional problems.

Lack of generalizability

The findings may not be generalizable to other types of neural networks or PDEs, requiring further investigation to establish broader applicability.

Expert Commentary

The article presents a comprehensive empirical analysis of single-layer PINNs on canonical nonlinear PDEs, highlighting the optimization challenges and limitations of these networks. The authors' findings, particularly the identification of dual optimization failures, shed light on the complexity of scaling behavior beyond separable power laws. The proposed methodology to empirically measure complex scaling effects is a valuable contribution to the field of PINNs. However, the study's restrictive scope and lack of generalizability limit its broader applicability. Future research should aim to extend these findings to more complex or multi-dimensional problems and explore the applicability of these results to other types of neural networks and PDEs.

Recommendations

  • Future research should focus on developing more sophisticated network architectures to tackle nonlinear PDEs.
  • The proposed methodology should be further validated and extended to other types of neural networks and PDEs to establish broader applicability.

Sources