Learning Effective Dynamics across Spatio-Temporal Scales of Complex Flows

📅 2025-02-11
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Addressing the challenge of modeling dynamic cross-scale spatiotemporal interactions in multiscale complex fluid dynamics, this paper proposes Graph-LED—a novel framework coupling graph neural networks (GNNs) with autoregressive attention mechanisms. It innovatively encodes unstructured-grid flow fields as dynamic graphs, synergistically leveraging GNNs’ spatial dimensionality reduction and attention mechanisms’ temporal modeling capability to enable efficient multiscale dynamical learning on variable-resolution meshes. The method requires only limited simulation data to capture cross-scale coupling relationships. Evaluated on canonical benchmarks—including flow past a circular cylinder and backward-facing step flow—Graph-LED accurately reproduces near-wall fine-scale structures and far-field wake evolution, significantly outperforming conventional reduced-order models. This work establishes a scalable, data-efficient modeling paradigm for complex flows where full-scale direct numerical simulation remains computationally intractable.

Technology Category

Application Category

📝 Abstract
Modeling and simulation of complex fluid flows with dynamics that span multiple spatio-temporal scales is a fundamental challenge in many scientific and engineering domains. Full-scale resolving simulations for systems such as highly turbulent flows are not feasible in the foreseeable future, and reduced-order models must capture dynamics that involve interactions across scales. In the present work, we propose a novel framework, Graph-based Learning of Effective Dynamics (Graph-LED), that leverages graph neural networks (GNNs), as well as an attention-based autoregressive model, to extract the effective dynamics from a small amount of simulation data. GNNs represent flow fields on unstructured meshes as graphs and effectively handle complex geometries and non-uniform grids. The proposed method combines a GNN based, dimensionality reduction for variable-size unstructured meshes with an autoregressive temporal attention model that can learn temporal dependencies automatically. We evaluated the proposed approach on a suite of fluid dynamics problems, including flow past a cylinder and flow over a backward-facing step over a range of Reynolds numbers. The results demonstrate robust and effective forecasting of spatio-temporal physics; in the case of the flow past a cylinder, both small-scale effects that occur close to the cylinder as well as its wake are accurately captured.
Problem

Research questions and friction points this paper is trying to address.

Modeling complex fluid flows
Learning cross-scale dynamics
Forecasting spatio-temporal physics
Innovation

Methods, ideas, or system contributions that make the work stand out.

Graph neural networks
Attention-based autoregressive model
Unstructured meshes dimensionality reduction
🔎 Similar Papers
No similar papers found.