SPUS: A Lightweight and Parameter-Efficient Foundation Model for PDEs

πŸ“… 2025-10-01
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Existing PDE foundation models rely heavily on large-scale Transformer architectures, incurring prohibitive parameter counts and computational costs. To address this, we propose SPUSβ€”the first unified neural operator foundation model built upon a lightweight residual U-Net architecture. Our key contributions are threefold: (i) the first integration of residual U-Net into PDE foundation modeling; (ii) a physics-informed autoregressive pretraining strategy that explicitly emulates numerical PDE solvers to learn conservation laws and dynamical evolution; and (iii) joint pretraining on multi-physics fluid PDEs coupled with few-shot transfer mechanisms. Experiments demonstrate that SPUS achieves state-of-the-art generalization across six unseen PDE tasks, reducing parameter count by one to two orders of magnitude compared to mainstream approaches. Moreover, it adapts rapidly to novel equations with only a few fine-tuning samples.

Technology Category

Application Category

πŸ“ Abstract
We introduce Small PDE U-Net Solver (SPUS), a compact and efficient foundation model (FM) designed as a unified neural operator for solving a wide range of partial differential equations (PDEs). Unlike existing state-of-the-art PDE FMs-primarily based on large complex transformer architectures with high computational and parameter overhead-SPUS leverages a lightweight residual U-Net-based architecture that has been largely underexplored as a foundation model architecture in this domain. To enable effective learning in this minimalist framework, we utilize a simple yet powerful auto-regressive pretraining strategy which closely replicates the behavior of numerical solvers to learn the underlying physics. SPUS is pretrained on a diverse set of fluid dynamics PDEs and evaluated across 6 challenging unseen downstream PDEs spanning various physical systems. Experimental results demonstrate that SPUS using residual U-Net based architecture achieves state-of-the-art generalization on these downstream tasks while requiring significantly fewer parameters and minimal fine-tuning data, highlighting its potential as a highly parameter-efficient FM for solving diverse PDE systems.
Problem

Research questions and friction points this paper is trying to address.

Develops lightweight neural operator for solving diverse PDEs
Uses efficient U-Net architecture to reduce computational overhead
Achieves parameter-efficient generalization across multiple physical systems
Innovation

Methods, ideas, or system contributions that make the work stand out.

Lightweight residual U-Net architecture for PDEs
Auto-regressive pretraining mimicking numerical solvers
Parameter-efficient foundation model with minimal fine-tuning
πŸ”Ž Similar Papers
No similar papers found.
A
Abu Bucker Siddik
Computing and Artificial Intelligence Division (CAI), Los Alamos National Laboratory, Los Alamos, New Mexico, US, 87545
Diane Oyen
Diane Oyen
Los Alamos National Laboratory
Machine Learning
A
Alexander Most
Computing and Artificial Intelligence Division (CAI), Los Alamos National Laboratory, Los Alamos, New Mexico, US, 87545
Michal Kucer
Michal Kucer
Staff Scientist, Los Alamos National Laboratory
Computer VisionDeep LearningMachine Learning
A
Ayan Biswas
Computing and Artificial Intelligence Division (CAI), Los Alamos National Laboratory, Los Alamos, New Mexico, US, 87545