🤖 AI Summary
In energy-harvesting federated learning (EHFL), client participation is unstable and computationally energy-intensive due to battery constraints. Method: This paper proposes FedBacys, a battery-aware cyclic scheduling framework, and its lightweight variant FedBacys-Odd. Its core innovation lies in the first integration of battery-state-driven periodic participation with clustering-guided selective client selection—reducing redundant computation while preserving model convergence. Contribution/Results: Theoretical analysis establishes convergence guarantees. Experiments demonstrate that, compared to state-of-the-art baselines, FedBacys achieves comparable model accuracy and convergence speed while reducing system energy consumption by 32.7%, significantly improving device participation robustness and energy utilization efficiency.
📝 Abstract
Federated Learning (FL) is a powerful paradigm for distributed learning, but its increasing complexity leads to significant energy consumption from client-side computations for training models. In particular, the challenge is critical in energy-harvesting FL (EHFL) systems where participation availability of each device oscillates due to limited energy. To address this, we propose FedBacys, a battery-aware EHFL framework using cyclic client participation based on users' battery levels. By clustering clients and scheduling them sequentially, FedBacys minimizes redundant computations, reduces system-wide energy usage, and improves learning stability. We also introduce FedBacys-Odd, a more energy-efficient variant that allows clients to participate selectively, further reducing energy costs without compromising performance. We provide a convergence analysis for our framework and demonstrate its superior energy efficiency and robustness compared to existing algorithms through numerical experiments.