Unlearning-Enhanced Website Fingerprinting Attack: Against Backdoor Poisoning in Anonymous Networks

📅 2025-06-16
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Website fingerprinting (WF) models in anonymous networks are highly vulnerable to backdoor poisoning attacks, leading to severe performance degradation. To address this, this work pioneers the application of machine unlearning to WF defense, proposing a lightweight backdoor detection and purification method based on influence functions. Our approach quantifies gradient sensitivity and dynamically reweights model parameters to automatically localize poisoned samples and perform parameter-level model purification—requiring only a small number of known poisoned instances to debias the training set and eliminate backdoors. Evaluated on the CW and OW public datasets, our method achieves stable poisoning and test accuracy of 80%, significantly outperforming baseline defenses. Moreover, it improves runtime efficiency by 2–3× while maintaining robustness, discriminative capability, and resilience against interference under open-world assumptions.

Technology Category

Application Category

📝 Abstract
Website Fingerprinting (WF) is an effective tool for regulating and governing the dark web. However, its performance can be significantly degraded by backdoor poisoning attacks in practical deployments. This paper aims to address the problem of hidden backdoor poisoning attacks faced by Website Fingerprinting attack, and designs a feasible mothed that integrates unlearning technology to realize detection of automatic poisoned points and complete removal of its destructive effects, requiring only a small number of known poisoned test points. Taking Tor onion routing as an example, our method evaluates the influence value of each training sample on these known poisoned test points as the basis for judgment. We optimize the use of influence scores to identify poisoned samples within the training dataset. Furthermore, by quantifying the difference between the contribution of model parameters on the taining data and the clean data, the target parameters are dynamically adjusted to eliminate the impact of the backdoor attacks. Experiments on public datasets under the assumptions of closed-world (CW) and open-world (OW) verify the effectiveness of the proposed method. In complex scenes containing both clean website fingerprinting features and backdoor triggers, the accuracy of the model on the poisoned dataset and the test dataset is stable at about 80%, significantly outperforming the traditional WF attack models. In addition, the proposed method achieves a 2-3 times speedup in runtime efficiency compared to baseline methods. By incorporating machine unlearning, we realize a WF attack model that exhibits enhanced resistance to backdoor poisoning and faster execution speeds in adversarial settings.
Problem

Research questions and friction points this paper is trying to address.

Detect and remove backdoor poisoning in Website Fingerprinting attacks
Enhance attack resistance using unlearning technology
Improve runtime efficiency in adversarial settings
Innovation

Methods, ideas, or system contributions that make the work stand out.

Integrates unlearning for backdoor attack resistance
Uses influence scores to detect poisoned samples
Dynamically adjusts parameters to eliminate backdoor effects
🔎 Similar Papers
No similar papers found.
Yali Yuan
Yali Yuan
University of Göttingen
Intelligent Internet of Thingsnetwork intrusion and attack detectionprivacy protectionnetwork localization and security
K
Kai Xu
School of Cyber Science and Engineering, Southeast University
R
Ruolin Ma
School of Cyber Science and Engineering, Southeast University
Y
Yuchen Zhang
School of Cyber Science and Engineering, Southeast University