🤖 AI Summary
To address GPU memory constraints and inter-device communication bottlenecks in distributed training of large neural networks, this paper proposes a segmented collaborative training framework based on synthetic labels: lightweight synthetic intermediate labels replace high-dimensional real feature transmissions, decoupling forward and backward propagation dependencies. The framework introduces a label generator, consistency regularization, and segmented asynchronous backpropagation to eliminate gradient synchronization overhead. Evaluated on ResNet-50/CIFAR-10, the method reduces communication volume by 92%, accelerates training by 2.3×, incurs <0.5% accuracy degradation, and scales losslessly to an 8-GPU cluster. This work is the first to systematically apply synthetic labels to cross-device collaborative training, significantly improving training efficiency and scalability for large-scale models.