🤖 AI Summary
Existing graph neural network (GNN) ensemble methods suffer from slow inference, high memory overhead, and poor scalability. To address these limitations, this paper proposes a gradient-descent-based learnable Soup strategy—introducing *Learned Souping* and its block-wise variant, *Partition Learned Souping*. Unlike conventional fixed-weight ensembling, our approach jointly optimizes model parameters and weighting coefficients in an end-to-end manner, achieving superior accuracy with significantly reduced computational and memory costs. On multiple OGB benchmarks, the method improves test accuracy by up to 1.2%, accelerates inference by 2.1× overall, and on ogbn-products, the Partition variant achieves a 24.5× speedup and 76% memory reduction—overcoming the efficiency bottlenecks inherent in standard Soup methods.
📝 Abstract
Graph Neural Networks (GNN) have demonstrated state-of-the-art performance in numerous scientific and high-performance computing (HPC) applications. Recent work suggests that"souping"(combining) individually trained GNNs into a single model can improve performance without increasing compute and memory costs during inference. However, existing souping algorithms are often slow and memory-intensive, which limits their scalability. We introduce Learned Souping for GNNs, a gradient-descent-based souping strategy that substantially reduces time and memory overhead compared to existing methods. Our approach is evaluated across multiple Open Graph Benchmark (OGB) datasets and GNN architectures, achieving up to 1.2% accuracy improvement and 2.1X speedup. Additionally, we propose Partition Learned Souping, a novel partition-based variant of learned souping that significantly reduces memory usage. On the ogbn-products dataset with GraphSAGE, partition learned souping achieves a 24.5X speedup and a 76% memory reduction without compromising accuracy.