🤖 AI Summary
Dynamic system reconstruction (DSR) suffers from weak cross-level generalization and struggles to effectively learn from sparse, heterogeneous, multi-scale dynamical data.
Method: This paper proposes a hierarchical meta-learning framework centered on MixER—a novel sparse Top-1 Mixture-of-Experts layer. Departing from gradient-based gating, MixER introduces a customized gating update mechanism jointly driven by K-means clustering and least-squares optimization, explicitly modeling the intrinsic hierarchical structure of data. The method integrates ordinary differential equation (ODE) modeling, sparse activation, and hierarchical meta-learning for efficient, scalable training.
Results: Evaluated on parameterized ODE systems with ≤10 dimensions, the framework demonstrates strong scalability. Experiments on synthetic benchmarks and neuroscience time-series data show significantly improved representation quality over baselines, with high dependence on—and accurate capture of—data hierarchy.
📝 Abstract
As foundational models reshape scientific discovery, a bottleneck persists in dynamical system reconstruction (DSR): the ability to learn across system hierarchies. Many meta-learning approaches have been applied successfully to single systems, but falter when confronted with sparse, loosely related datasets requiring multiple hierarchies to be learned. Mixture of Experts (MoE) offers a natural paradigm to address these challenges. Despite their potential, we demonstrate that naive MoEs are inadequate for the nuanced demands of hierarchical DSR, largely due to their gradient descent-based gating update mechanism which leads to slow updates and conflicted routing during training. To overcome this limitation, we introduce MixER: Mixture of Expert Reconstructors, a novel sparse top-1 MoE layer employing a custom gating update algorithm based on $K$-means and least squares. Extensive experiments validate MixER's capabilities, demonstrating efficient training and scalability to systems of up to ten parametric ordinary differential equations. However, our layer underperforms state-of-the-art meta-learners in high-data regimes, particularly when each expert is constrained to process only a fraction of a dataset composed of highly related data points. Further analysis with synthetic and neuroscientific time series suggests that the quality of the contextual representations generated by MixER is closely linked to the presence of hierarchical structure in the data.