Ensemble and Mixture-of-Experts DeepONets For Operator Learning

πŸ“… 2024-05-20
πŸ›οΈ arXiv.org
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
To address the limited expressivity and generalization capability of DeepONet in learning operators for two- and three-dimensional partial differential equations (PDEs), this paper proposes PoU-MoE DeepONetβ€”a novel multi-trunk ensemble architecture that integrates partition-of-unity (PoU)-based sparse modeling with a spatial mixture-of-experts (MoE) mechanism, systematically augmented by basis enhancement and POD-guided spatial localization. We establish its universal approximation property theoretically. On multiple 2D/3D PDE operator learning benchmarks, PoU-MoE DeepONet achieves 2–4Γ— lower relative β„“β‚‚ error compared to standard DeepONet and POD-DeepONet, significantly improving accuracy, robustness, and spatially localized representation. The framework offers a scalable and interpretable paradigm for learning complex physical operators.

Technology Category

Application Category

πŸ“ Abstract
We present a novel deep operator network (DeepONet) architecture for operator learning, the ensemble DeepONet, that allows for enriching the trunk network of a single DeepONet with multiple distinct trunk networks. This trunk enrichment allows for greater expressivity and generalization capabilities over a range of operator learning problems. We also present a spatial mixture-of-experts (MoE) DeepONet trunk network architecture that utilizes a partition-of-unity (PoU) approximation to promote spatial locality and model sparsity in the operator learning problem. We first prove that both the ensemble and PoU-MoE DeepONets are universal approximators. We then demonstrate that ensemble DeepONets containing a trunk ensemble of a standard trunk, the PoU-MoE trunk, and/or a proper orthogonal decomposition (POD) trunk can achieve 2-4x lower relative $ell_2$ errors than standard DeepONets and POD-DeepONets on both standard and challenging new operator learning problems involving partial differential equations (PDEs) in two and three dimensions. Our new PoU-MoE formulation provides a natural way to incorporate spatial locality and model sparsity into any neural network architecture, while our new ensemble DeepONet provides a powerful and general framework for incorporating basis enrichment in scientific machine learning architectures for operator learning.
Problem

Research questions and friction points this paper is trying to address.

Enhance operator learning with ensemble DeepONet architecture
Introduce spatial mixture-of-experts for model sparsity and locality
Achieve lower errors in PDE-related operator learning problems
Innovation

Methods, ideas, or system contributions that make the work stand out.

Ensemble DeepONet enhances trunk network expressivity.
Spatial MoE DeepONet promotes locality and sparsity.
Combines multiple trunk types for improved error rates.
πŸ”Ž Similar Papers
No similar papers found.
R
Ramansh Sharma
Kahlert School of Computing, University of Utah, UT, USA
Varun Shankar
Varun Shankar
Kahlert School of Computing, University of Utah
Scientific Machine LearningScientific Computing