🤖 AI Summary
Neural ordinary differential equations (Neural ODEs) suffer from high computational overhead due to repeated numerical integration during training. To address this, we propose a fully discretized co-optimization framework. Our method reformulates the continuous dynamics and parameter learning as a large-scale nonlinear programming problem via collocation-based full discretization—a novel simultaneous optimization paradigm. We further introduce an alternating direction method of multipliers (ADMM) decomposition strategy to enable distributed, inter-batch co-optimization of submodels. The discretized optimization problem is efficiently solved using the IPOPT solver. Evaluated on the Van der Pol oscillator benchmark, our approach achieves significantly faster convergence and higher training efficiency compared to standard Neural ODE training. Experimental results demonstrate superior computational efficiency, numerical stability, and scalability—validating the framework’s effectiveness for optimizing continuous-depth models without sacrificing accuracy or robustness.
📝 Abstract
Neural Ordinary Differential Equations (Neural ODEs) represent continuous-time dynamics with neural networks, offering advancements for modeling and control tasks. However, training Neural ODEs requires solving differential equations at each epoch, leading to high computational costs. This work investigates simultaneous optimization methods as a faster training alternative. In particular, we employ a collocation-based, fully discretized formulation and use IPOPT--a solver for large-scale nonlinear optimization--to simultaneously optimize collocation coefficients and neural network parameters. Using the Van der Pol Oscillator as a case study, we demonstrate faster convergence compared to traditional training methods. Furthermore, we introduce a decomposition framework utilizing Alternating Direction Method of Multipliers (ADMM) to effectively coordinate sub-models among data batches. Our results show significant potential for (collocation-based) simultaneous Neural ODE training pipelines.