๐ค AI Summary
Recent low-communication training advances have blurred conceptual distinctions between multi-cluster distributed training (e.g., cross-datacenter coordination) and community-driven decentralized training (e.g., P2P collaboration), leading to conflation in policy discourse. Method: This work systematically differentiates these paradigms, introducing a computationally traceable analytical framework that integrates distributed optimization theory with decentralized training architectures. Contribution/Results: It reveals divergent implications across three governance dimensions: (i) computational infrastructure risk (increased opacity), (ii) capability diffusion risk (lowered entry barriers), and (iii) model detectability and shutdown feasibility (eroded regulatory control). Findings affirm the continued relevance of export controls while highlighting decentralized trainingโs potential benefits for privacy preservation and antitrust objectives. The study provides both theoretical foundations and technical pathways for precision policymaking in compute governance, capability containment, and infrastructure oversight.
๐ Abstract
Advances in low-communication training algorithms are enabling a shift from centralised model training to compute setups that are either distributed across multiple clusters or decentralised via community-driven contributions. This paper distinguishes these two scenarios - distributed and decentralised training - which are little understood and often conflated in policy discourse. We discuss how they could impact technical AI governance through an increased risk of compute structuring, capability proliferation, and the erosion of detectability and shutdownability. While these trends foreshadow a possible new paradigm that could challenge key assumptions of compute governance, we emphasise that certain policy levers, like export controls, remain relevant. We also acknowledge potential benefits of decentralised AI, including privacy-preserving training runs that could unlock access to more data, and mitigating harmful power concentration. Our goal is to support more precise policymaking around compute, capability proliferation, and decentralised AI development.