FedBAP: Backdoor Defense via Benign Adversarial Perturbation in Federated Learning

πŸ“… 2025-07-26
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Federated learning (FL) is vulnerable to backdoor attacks, and existing defenses are limited because they neglect models’ over-reliance on triggers. To address this, we propose a novel defense framework based on benign adversarial perturbations. First, we design a perturbation-triggering mechanism that generates benign adversarial examples semantically aligned with the backdoor. Second, we introduce an adaptive scaling mechanism that dynamically modulates perturbation intensity to suppress backdoor dependency while preserving model accuracy. Our method integrates seamlessly into the FL training pipeline without modifying clients’ local update procedures. Extensive experiments demonstrate that it reduces attack success rates by 0.22–5.34%, 0.48–6.34%, and 97.22–97.6% against three representative backdoor attacks, respectively. Moreover, it exhibits strong generalization against unseen backdoor patterns. Overall, our approach significantly enhances the robustness of FL systems against backdoor threats.

Technology Category

Application Category

πŸ“ Abstract
Federated Learning (FL) enables collaborative model training while preserving data privacy, but it is highly vulnerable to backdoor attacks. Most existing defense methods in FL have limited effectiveness due to their neglect of the model's over-reliance on backdoor triggers, particularly as the proportion of malicious clients increases. In this paper, we propose FedBAP, a novel defense framework for mitigating backdoor attacks in FL by reducing the model's reliance on backdoor triggers. Specifically, first, we propose a perturbed trigger generation mechanism that creates perturbation triggers precisely matching backdoor triggers in location and size, ensuring strong influence on model outputs. Second, we utilize these perturbation triggers to generate benign adversarial perturbations that disrupt the model's dependence on backdoor triggers while forcing it to learn more robust decision boundaries. Finally, we design an adaptive scaling mechanism to dynamically adjust perturbation intensity, effectively balancing defense strength and model performance. The experimental results demonstrate that FedBAP reduces the attack success rates by 0.22%-5.34%, 0.48%-6.34%, and 97.22%-97.6% under three types of backdoor attacks, respectively. In particular, FedBAP demonstrates outstanding performance against novel backdoor attacks.
Problem

Research questions and friction points this paper is trying to address.

Defends against backdoor attacks in Federated Learning
Reduces model reliance on backdoor triggers
Balances defense strength and model performance
Innovation

Methods, ideas, or system contributions that make the work stand out.

Generates perturbed triggers matching backdoor triggers
Uses benign adversarial perturbations for robust learning
Adaptively scales perturbation intensity dynamically
X
Xinhai Yan
School of Cyber Science and Engineering, Wuhan University
Libing Wu
Libing Wu
wuhan university
Z
Zhuangzhuang Zhang
School of Cyber Science and Engineering, Wuhan University
Bingyi Liu
Bingyi Liu
Professor, Department of CS and AI, Wuhan University of Technology
Internet of VehiclesEdge ComputingAutonomous VehiclesIntelligent Transportation Systems
L
Lijuan Huo
School of Cyber Science and Engineering, Wuhan University
J
Jing Wang
School of Software Engineering, Huazhong University of Science and Technology