🤖 AI Summary
This study systematically investigates the practical utility and applicability boundaries of global attention mechanisms in atomic graph learning, addressing ambiguity in reported contributions arising from inconsistent implementations and hyperparameter tuning. We propose the first reproducible, modularly decoupled hybrid model benchmark framework—built upon HydraGNN—that orthogonally integrates message-passing neural networks (MPNNs), global attention, and chemical or topological encoders. Evaluated across seven diverse datasets on both regression and classification tasks, our results demonstrate that encoder-augmented MPNNs establish a strong baseline; models featuring deep local–global fusion significantly outperform others on tasks dominated by long-range interactions; and we provide the first quantitative characterization of the trade-off among accuracy gains, computational overhead, and memory cost introduced by attention mechanisms. The framework enables rigorous, controlled ablation studies and facilitates fair comparison across architectural choices in atomic graph representation learning.
📝 Abstract
Graph neural networks (GNNs) are widely used as surrogates for costly experiments and first-principles simulations to study the behavior of compounds at atomistic scale, and their architectural complexity is constantly increasing to enable the modeling of complex physics. While most recent GNNs combine more traditional message passing neural networks (MPNNs) layers to model short-range interactions with more advanced graph transformers (GTs) with global attention mechanisms to model long-range interactions, it is still unclear when global attention mechanisms provide real benefits over well-tuned MPNN layers due to inconsistent implementations, features, or hyperparameter tuning. We introduce the first unified, reproducible benchmarking framework - built on HydraGNN - that enables seamless switching among four controlled model classes: MPNN, MPNN with chemistry/topology encoders, GPS-style hybrids of MPNN with global attention, and fully fused local - global models with encoders. Using seven diverse open-source datasets for benchmarking across regression and classification tasks, we systematically isolate the contributions of message passing, global attention, and encoder-based feature augmentation. Our study shows that encoder-augmented MPNNs form a robust baseline, while fused local-global models yield the clearest benefits for properties governed by long-range interaction effects. We further quantify the accuracy - compute trade-offs of attention, reporting its overhead in memory. Together, these results establish the first controlled evaluation of global attention in atomistic graph learning and provide a reproducible testbed for future model development.