Dual-View Inference Attack: Machine Unlearning Amplifies Privacy Exposure

📅 2025-12-17
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work uncovers a novel privacy risk induced by machine unlearning in the dual-view setting (original model + unlearned model): while protecting the requested-forget samples, it significantly exacerbates membership inference attacks (MIAs) against retained training data. We propose the Dual-View Inference Attack (DVIA), the first attack paradigm that theoretically formalizes this threat from an information-theoretic perspective—defining “privacy knowledge gain” to quantify the amplification of privacy leakage under joint querying. DVIA requires no auxiliary attack model training and operates in a black-box, lightweight, and cross-architecture manner. Extensive experiments across multiple datasets and model architectures demonstrate that DVIA boosts MIA accuracy by up to 32.7%, revealing that machine unlearning may inadvertently weaken—rather than strengthen—the overall privacy protection of training data.

Technology Category

Application Category

📝 Abstract
Machine unlearning is a newly popularized technique for removing specific training data from a trained model, enabling it to comply with data deletion requests. While it protects the rights of users requesting unlearning, it also introduces new privacy risks. Prior works have primarily focused on the privacy of data that has been unlearned, while the risks to retained data remain largely unexplored. To address this gap, we focus on the privacy risks of retained data and, for the first time, reveal the vulnerabilities introduced by machine unlearning under the dual-view setting, where an adversary can query both the original and the unlearned models. From an information-theoretic perspective, we introduce the concept of {privacy knowledge gain} and demonstrate that the dual-view setting allows adversaries to obtain more information than querying either model alone, thereby amplifying privacy leakage. To effectively demonstrate this threat, we propose DVIA, a Dual-View Inference Attack, which extracts membership information on retained data using black-box queries to both models. DVIA eliminates the need to train an attack model and employs a lightweight likelihood ratio inference module for efficient inference. Experiments across different datasets and model architectures validate the effectiveness of DVIA and highlight the privacy risks inherent in the dual-view setting.
Problem

Research questions and friction points this paper is trying to address.

Investigates privacy risks to retained data after machine unlearning.
Reveals vulnerabilities from querying both original and unlearned models.
Proposes an attack extracting membership info without training an attack model.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Dual-view attack queries both original and unlearned models
Privacy knowledge gain measured via information-theoretic analysis
Lightweight likelihood ratio inference without training attack model
🔎 Similar Papers
No similar papers found.
L
Lulu Xue
School of Cyber Science and Engineering, Huazhong University of Science and Technology
Shengshan Hu
Shengshan Hu
School of CSE, Huazhong University of Science and Technology (HUST)
AI SecurityEmbodied AIAutonomous Driving
L
Linqiang Qian
Institute for Network Sciences and Cyberspace, Tsinghua University
P
Peijin Guo
School of Cyber Science and Engineering, Huazhong University of Science and Technology
Y
Yechao Zhang
College of Computing and Data Science, Nanyang Technological University
Minghui Li
Minghui Li
Huazhong University of Science and Technology
AI Security
Yanjun Zhang
Yanjun Zhang
Lecturer, University of Technology Sydney
Security and PrivacyMachine Learning
Dayong Ye
Dayong Ye
University of Technology, Sydney
Multi-Agent SystemsPrivacy Preservation
L
Leo Yu Zhang
School of Information and Communication Technology, Griffith University