🤖 AI Summary
Distributed stochastic optimization faces arbitrary computational dynamics—including hardware disconnections, time-varying compute capacity, and fluctuating processing speeds—rendering existing models inadequate for real-world deployment.
Method: We propose the first general asynchronous computation model encompassing all realistic scenarios; based on it, we derive tight time-complexity lower bounds (up to constant factors) for mainstream synchronous and asynchronous methods—including Minibatch SGD, Async SGD, and Picky SGD—and prove that Rennala/Malenia SGD achieves optimal convergence. Our analysis integrates general computational dynamical modeling, information-theoretic lower-bound derivation, and stochastic optimization convergence theory.
Contribution: We establish fundamental theoretical limits and design principles for system-aware optimization, providing foundational support for fault-tolerant, robust distributed learning. The results unify treatment of heterogeneous, unreliable, and dynamic execution environments while delivering precise complexity characterizations grounded in both system behavior and statistical learning theory.
📝 Abstract
In distributed stochastic optimization, where parallel and asynchronous methods are employed, we establish optimal time complexities under virtually any computation behavior of workers/devices/CPUs/GPUs, capturing potential disconnections due to hardware and network delays, time-varying computation powers, and any possible fluctuations and trends of computation speeds. These real-world scenarios are formalized by our new universal computation model. Leveraging this model and new proof techniques, we discover tight lower bounds that apply to virtually all synchronous and asynchronous methods, including Minibatch SGD, Asynchronous SGD (Recht et al., 2011), and Picky SGD (Cohen et al., 2021). We show that these lower bounds, up to constant factors, are matched by the optimal Rennala SGD and Malenia SGD methods (Tyurin&Richt'arik, 2023).