š¤ AI Summary
To address the violation of physical first principlesāleading to unphysical energy growth and instabilityāin neural dynamical models during long-term evolution, this paper proposes the Modular ConservativeāDissipative Modeling (MCFM) framework, which explicitly embeds the conservativeādissipative decomposition into both vector-field modeling and numerical sampling. Methodologically, MCFM introduces a symmetric-proximal discrete update scheme, combining symplectic integration with proximal-metric steps via Strang splitting, optionally augmented with energy projection to enforce monotonic energy decay in both continuous and discrete time. Theoretically, this work provides the first rigorous stability guarantee for such structured neural ODEs. Empirically, on mechanical system benchmarks, MCFM achieves significantly improved phase-space trajectory fidelity, drastically reduces spurious energy growth events, and matches the terminal distribution accuracy of unconstrained baselines.
š Abstract
Metriplectic conditional flow matching (MCFM) learns dissipative dynamics without violating first principles. Neural surrogates often inject energy and destabilize long-horizon rollouts; MCFM instead builds the conservative-dissipative split into both the vector field and a structure preserving sampler. MCFM trains via conditional flow matching on short transitions, avoiding long rollout adjoints. In inference, a Strang-prox scheme alternates a symplectic update with a proximal metric step, ensuring discrete energy decay; an optional projection enforces strict decay when a trusted energy is available. We provide continuous and discrete time guarantees linking this parameterization and sampler to conservation, monotonic dissipation, and stable rollouts. On a controlled mechanical benchmark, MCFM yields phase portraits closer to ground truth and markedly fewer energy-increase and positive energy rate events than an equally expressive unconstrained neural flow, while matching terminal distributional fit.