Communication-Efficient Split Learning via Adaptive Feature-Wise Compression

πŸ“… 2023-07-20
πŸ›οΈ arXiv.org
πŸ“ˆ Citations: 4
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
To address the excessive communication overhead of intermediate features and gradients in Split Learning (SL) for edge-based collaborative training, this paper proposes SplitFCβ€”a novel framework introducing the first dual-strategy adaptive compression mechanism based on column-wise vector dispersion. Specifically: (1) a standard-deviation-driven feature-level Dropout dynamically prunes low-informativeness channels; and (2) a range-driven adaptive quantization scheme, solved in closed form for optimal quantization levels, minimizes reconstruction error. Additionally, chained gradient clipping ensures convergence stability. Extensive experiments on MNIST, CIFAR-100, and CelebA demonstrate that SplitFC reduces communication volume by up to 87% while maintaining accuracy comparable to state-of-the-art SL methods. The framework establishes a new paradigm for efficient, communication-aware collaborative training at the edge.
πŸ“ Abstract
This paper proposes a novel communication-efficient split learning (SL) framework, named SplitFC, which reduces the communication overhead required for transmitting intermediate feature and gradient vectors during the SL training process. The key idea of SplitFC is to leverage different dispersion degrees exhibited in the columns of the matrices. SplitFC incorporates two compression strategies: (i) adaptive feature-wise dropout and (ii) adaptive feature-wise quantization. In the first strategy, the intermediate feature vectors are dropped with adaptive dropout probabilities determined based on the standard deviation of these vectors. Then, by the chain rule, the intermediate gradient vectors associated with the dropped feature vectors are also dropped. In the second strategy, the non-dropped intermediate feature and gradient vectors are quantized using adaptive quantization levels determined based on the ranges of the vectors. To minimize the quantization error, the optimal quantization levels of this strategy are derived in a closed-form expression. Simulation results on the MNIST, CIFAR-100, and CelebA datasets demonstrate that SplitFC outperforms state-of-the-art SL frameworks by significantly reducing communication overheads while maintaining high accuracy.
Problem

Research questions and friction points this paper is trying to address.

Reduces communication overhead in split learning
Compresses intermediate feature and gradient vectors
Maintains high accuracy while minimizing data transmission
Innovation

Methods, ideas, or system contributions that make the work stand out.

Adaptive feature-wise dropout strategy
Adaptive feature-wise quantization method
Closed-form optimal quantization levels derivation
πŸ”Ž Similar Papers
No similar papers found.
Y
Yong-Nam Oh
Department of Electrical Engineering, POSTECH, Pohang, Gyeongbuk 37673, Republic of Korea
J
Jaeho Lee
Department of Electrical Engineering, POSTECH, Pohang, Gyeongbuk 37673, Republic of Korea
Christopher G. Brinton
Christopher G. Brinton
Elmore Associate Professor of ECE, Purdue University
NetworkingMachine LearningCommunicationsEdge ComputingNextG Wireless
Yo-Seb Jeon
Yo-Seb Jeon
Associate Professor, EE, POSTECH
Wireless CommunicationsMachine LearningSignal Processing