FloE: On-the-Fly MoE Inference

📅 2025-05-09
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the high inference latency of Mixture-of-Experts (MoE) models on memory-constrained devices, caused by excessive expert parameter activation and PCIe bandwidth saturation. We propose a synergistic mechanism combining expert-level fine-grained dynamic compression and low-overhead sparse activation prediction. Our approach integrates parameter block-wise quantization, structured sparse encoding, a lightweight expert activation predictor, and zero-copy CPU-GPU streaming loading—achieving, for the first time, joint optimization of dynamic compression and activation prediction at the expert granularity. On Mixtral-8×7B, our method achieves a per-expert compression ratio of 9.3×, enabling deployment with only 11 GB GPU memory. Compared to DeepSpeed-MII, it improves end-to-end inference throughput by 48.7× and reduces peak GPU memory consumption by up to 8.5%. This work breaks the strong coupling between bandwidth and memory bottlenecks in MoE inference.

Technology Category

Application Category

📝 Abstract
With the widespread adoption of Mixture-of-Experts (MoE) models, there is a growing demand for efficient inference on memory-constrained devices. While offloading expert parameters to CPU memory and loading activated experts on demand has emerged as a potential solution, the large size of activated experts overburdens the limited PCIe bandwidth, hindering the effectiveness in latency-sensitive scenarios. To mitigate this, we propose FloE, an on-the-fly MoE inference system on memory-constrained GPUs. FloE is built on the insight that there exists substantial untapped redundancy within sparsely activated experts. It employs various compression techniques on the expert's internal parameter matrices to reduce the data movement load, combined with low-cost sparse prediction, achieving perceptible inference acceleration in wall-clock time on resource-constrained devices. Empirically, FloE achieves a 9.3x compression of parameters per expert in Mixtral-8x7B; enables deployment on a GPU with only 11GB VRAM, reducing the memory footprint by up to 8.5x; and delivers a 48.7x inference speedup compared to DeepSpeed-MII on a single GeForce RTX 3090.
Problem

Research questions and friction points this paper is trying to address.

Efficient MoE inference on memory-constrained devices
Reducing PCIe bandwidth overload from large expert activations
Compressing expert parameters to accelerate resource-limited inference
Innovation

Methods, ideas, or system contributions that make the work stand out.

Compresses expert parameters to reduce data movement
Enables MoE inference on memory-constrained GPUs
Combines compression with low-cost sparse prediction
🔎 Similar Papers
No similar papers found.
Yuxin Zhou
Yuxin Zhou
University of California, Riverside
CombustionNanoparticlesMolecular dynamicsAerosol
Z
Zheng Li
The State Key Laboratory of Blockchain and Data Security, Zhejiang University; Hangzhou High-Tech Zone (Binjiang) Institute of Blockchain and Data Security
J
Jun Zhang
The State Key Laboratory of Blockchain and Data Security, Zhejiang University; Hangzhou High-Tech Zone (Binjiang) Institute of Blockchain and Data Security
J
Jue Wang
The State Key Laboratory of Blockchain and Data Security, Zhejiang University; Hangzhou High-Tech Zone (Binjiang) Institute of Blockchain and Data Security
Y
Yiping Wang
Paul G. Allen School of Computer Science & Engineering, University of Washington
Zhongle Xie
Zhongle Xie
Zhejiang University
AI4DBML SystemsDB SystemsOLAP
K
Ke Chen
The State Key Laboratory of Blockchain and Data Security, Zhejiang University; Hangzhou High-Tech Zone (Binjiang) Institute of Blockchain and Data Security
Lidan Shou
Lidan Shou
Professor of Computer Science, Zhejiang University
DatabaseData & Knowledge ManagementML Systems