🤖 AI Summary
Large language model (LLM) inference in heterogeneous GPU clusters faces conflicting challenges of hardware resource heterogeneity and stringent low-latency requirements.
Method: This paper proposes a unified joint modeling framework that formalizes both model deployment and request scheduling as a maximum flow problem on a directed weighted graph, solved globally via mixed-integer linear programming (MILP). Crucially, GPUs serve as graph nodes, while network bandwidth and computational capacity constitute edge weights; graph representation learning is integrated to enable resource-aware scheduling.
Contribution/Results: The approach overcomes the performance bottlenecks of conventional decoupled strategies. Evaluated on real-world heterogeneous clusters comprising 24–42 GPUs, it achieves up to a 3.3× throughput improvement, a 66% reduction in time-to-first-token latency, and a 24% reduction in decoding latency.
📝 Abstract
This paper introduces Helix, a distributed system for high-throughput, low-latency large language model (LLM) serving in heterogeneous GPU clusters. The key idea behind Helix is to formulate inference computation of LLMs over heterogeneous GPUs and network connections as a max-flow problem on directed, weighted graphs, whose nodes represent GPU instances and edges capture both GPU and network heterogeneity through their capacities. Helix then uses a mixed integer linear programming (MILP) algorithm to discover highly optimized strategies to serve LLMs on heterogeneous GPUs. This approach allows Helix to jointly optimize model placement and request scheduling, two highly entangled tasks in heterogeneous LLM serving. Our evaluation on several heterogeneous clusters ranging from 24 to 42 GPU nodes shows that Helix improves serving throughput by up to 3.3x and reduces prompting and decoding latency by up to 66% and 24%, respectively, compared to existing approaches. Helix is available at https://github.com/Thesys-lab/Helix-ASPLOS25.