PTL-PINNs: Perturbation-Guided Transfer Learning with Physics- Informed Neural Networks for Nonlinear Systems

📅 2026-01-17
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work proposes a novel framework, termed Perturbation Theory–enhanced Physics-Informed Neural Networks (PTL-PINNs), which integrates perturbation theory with transfer learning to address the poor generalization and high computational cost of conventional PINNs when solving nonlinear differential equations. For the first time, perturbation methods are incorporated into the PINN architecture, enabling the analytical derivation of closed-form expressions from an approximately linearized perturbed system to replace portions of the gradient-based optimization. This approach substantially reduces computational complexity while enhancing generalization efficiency. Experimental results across multiple nonlinear dynamical systems demonstrate that PTL-PINNs achieve accuracy comparable to that of classical Runge–Kutta methods, with nearly an order-of-magnitude improvement in computational speed.

Technology Category

Application Category

📝 Abstract
Accurately and efficiently solving nonlinear differential equations is crucial for modeling dynamic behavior across science and engineering. Physics-Informed Neural Networks (PINNs) have emerged as a powerful solution that embeds physical laws in training by enforcing equation residuals. However, these struggle to model nonlinear dynamics, suffering from limited generalization across problems and long training times. To address these limitations, we propose a perturbation-guided transfer learning framework for PINNs (PTL-PINN), which integrates perturbation theory with transfer learning to efficiently solve nonlinear equations. Unlike gradient-based transfer learning, PTL-PINNs solve an approximate linear perturbative system using closed-form expressions, enabling rapid generalization with the time complexity of matrix-vector multiplication. We show that PTL-PINNs achieve accuracy comparable to various Runge-Kutta methods, with computational speeds up to one order of magnitude faster. To benchmark performance, we solve a broad set of problems, including nonlinear oscillators across various damping regimes, the equilibrium-centered Lotka-Volterra system, the KPP-Fisher and the Wave equation. Since perturbation theory sets the accuracy bound of PTL-PINNs, we systematically evaluate its practical applicability. This work connects long-standing perturbation methods with PINNs, demonstrating how perturbation theory can guide foundational models to solve nonlinear systems with speeds comparable to those of classical solvers.
Problem

Research questions and friction points this paper is trying to address.

nonlinear differential equations
Physics-Informed Neural Networks
generalization
training efficiency
nonlinear dynamics
Innovation

Methods, ideas, or system contributions that make the work stand out.

Physics-Informed Neural Networks
Perturbation Theory
Transfer Learning
Nonlinear Differential Equations
Computational Efficiency
🔎 Similar Papers
No similar papers found.