Weight-Parameterization in Continuous Time Deep Neural Networks for Surrogate Modeling

📅 2025-07-29
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the challenge of jointly achieving high expressivity, numerical stability, and computational efficiency in time-varying weight modeling for continuous-time deep learning (e.g., neural ODEs), this paper proposes a low-dimensional parameterization method based on Legendre orthogonal polynomials. Specifically, time-varying weights are constrained to a subspace spanned by a finite number of Legendre basis functions. This formulation preserves strong representational capacity while significantly improving training stability and reducing memory and computational overhead. Integrated with the ODE-ResNet architecture, the approach supports both discrete-then-optimize and optimize-then-discrete training paradigms. Empirical evaluation on three high-dimensional physics benchmark tasks demonstrates that, compared to unconstrained models and monomial-basis parameterizations, our method achieves comparable or superior accuracy while accelerating training convergence and reducing average computational cost by 30–50%.

Technology Category

Application Category

📝 Abstract
Continuous-time deep learning models, such as neural ordinary differential equations (ODEs), offer a promising framework for surrogate modeling of complex physical systems. A central challenge in training these models lies in learning expressive yet stable time-varying weights, particularly under computational constraints. This work investigates weight parameterization strategies that constrain the temporal evolution of weights to a low-dimensional subspace spanned by polynomial basis functions. We evaluate both monomial and Legendre polynomial bases within neural ODE and residual network (ResNet) architectures under discretize-then-optimize and optimize-then-discretize training paradigms. Experimental results across three high-dimensional benchmark problems show that Legendre parameterizations yield more stable training dynamics, reduce computational cost, and achieve accuracy comparable to or better than both monomial parameterizations and unconstrained weight models. These findings elucidate the role of basis choice in time-dependent weight parameterization and demonstrate that using orthogonal polynomial bases offers a favorable tradeoff between model expressivity and training efficiency.
Problem

Research questions and friction points this paper is trying to address.

Learning stable time-varying weights in neural ODEs
Reducing computational cost in surrogate modeling
Improving training dynamics with orthogonal polynomial bases
Innovation

Methods, ideas, or system contributions that make the work stand out.

Polynomial basis functions constrain weight evolution
Legendre bases enhance training stability and efficiency
Orthogonal bases balance expressivity and computational cost
🔎 Similar Papers
No similar papers found.