Weighted Loss Methods for Robust Federated Learning under Data Heterogeneity

📅 2025-06-11
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
In federated learning under data heterogeneity, distinguishing honest anomalous gradients from Byzantine malicious ones remains challenging. To address this, we propose the Worker Label Alignment Loss (WoLA), a novel weighted loss function that— for the first time—incorporates label alignment into gradient direction modeling to explicitly enhance consistency among honest gradients across heterogeneous clients. Our method integrates WoLA-based optimization, robust aggregation, and rigorous theoretical convergence analysis. Evaluated on multiple heterogeneous benchmarks, it achieves >95% Byzantine attack detection success rate and improves convergence stability by over 40%, significantly outperforming existing state-of-the-art approaches. Furthermore, we provide formal theoretical guarantees on both convergence and robustness under realistic heterogeneity and adversarial assumptions.

Technology Category

Application Category

📝 Abstract
Federated learning (FL) is a machine learning paradigm that enables multiple data holders to collaboratively train a machine learning model without sharing their training data with external parties. In this paradigm, workers locally update a model and share with a central server their updated gradients (or model parameters). While FL seems appealing from a privacy perspective, it opens a number of threats from a security perspective as (Byzantine) participants can contribute poisonous gradients (or model parameters) harming model convergence. Byzantine-resilient FL addresses this issue by ensuring that the training proceeds as if Byzantine participants were absent. Towards this purpose, common strategies ignore outlier gradients during model aggregation, assuming that Byzantine gradients deviate more from honest gradients than honest gradients do from each other. However, in heterogeneous settings, honest gradients may differ significantly, making it difficult to distinguish honest outliers from Byzantine ones. In this paper, we introduce the Worker Label Alignement Loss (WoLA), a weighted loss that aligns honest worker gradients despite data heterogeneity, which facilitates the identification of Byzantines' gradients. This approach significantly outperforms state-of-the-art methods in heterogeneous settings. In this paper, we provide both theoretical insights and empirical evidence of its effectiveness.
Problem

Research questions and friction points this paper is trying to address.

Distinguishing Byzantine gradients from honest ones in federated learning
Addressing data heterogeneity in federated learning robustness
Improving model convergence with weighted loss methods
Innovation

Methods, ideas, or system contributions that make the work stand out.

Weighted loss aligns honest worker gradients
WoLA facilitates Byzantine gradient identification
Outperforms state-of-the-art in heterogeneous settings
🔎 Similar Papers
No similar papers found.
J
Johan Erbani
INSA Lyon, CNRS, UCBL, LIRIS, UMR 5205, Université de Lyon, 69621 Villeurbanne, France
Sonia Ben Mokhtar
Sonia Ben Mokhtar
LIRIS CNRS
Distributed systemsFault tolerancePrivacyDistributed Machine Learning
P
Pierre-Edouard Portier
Caisse d’Epargne Rhône-Alpes, Tour Incity, 116 Cours Lafayette, 69003 Lyon
E
Elod Egyed-Zsigmond
INSA Lyon, CNRS, UCBL, LIRIS, UMR 5205, Université de Lyon, 69621 Villeurbanne, France
D
Diana Nurbakova
INSA Lyon, CNRS, UCBL, LIRIS, UMR 5205, Université de Lyon, 69621 Villeurbanne, France