Visual Privacy Auditing with Diffusion Models

📅 2024-03-12
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the insufficient privacy guarantees of differential privacy (DP) image defenses in practical settings. We propose a diffusion-model (DM)-based data reconstruction attack framework that assesses visual privacy risks solely using real-image priors—without requiring access to the target model or training data. Notably, we are the first to repurpose Stable Diffusion as a visual privacy auditing tool, revealing the critical role of image priors in reconstruction success. Empirical evaluation demonstrates that standard DP theoretical bounds severely underestimate actual visual privacy leakage on CIFAR-10 and CelebA. Our contributions include: (1) establishing the first empirical privacy auditing benchmark tailored to visual data; (2) generating interpretable privacy leakage heatmaps that localize vulnerable image regions; and (3) exposing fundamental limitations of DP theoretical guarantees, thereby providing data-driven guidance for selecting DP hyperparameters in vision tasks.

Technology Category

Application Category

📝 Abstract
Data reconstruction attacks on machine learning models pose a substantial threat to privacy, potentially leaking sensitive information. Although defending against such attacks using differential privacy (DP) provides theoretical guarantees, determining appropriate DP parameters remains challenging. Current formal guarantees on the success of data reconstruction suffer from overly stringent assumptions regarding adversary knowledge about the target data, particularly in the image domain, raising questions about their real-world applicability. In this work, we empirically investigate this discrepancy by introducing a reconstruction attack based on diffusion models (DMs) that only assumes adversary access to real-world image priors and specifically targets the DP defense. We find that (1) real-world data priors significantly influence reconstruction success, (2) current reconstruction bounds do not model the risk posed by data priors well, and (3) DMs can serve as heuristic auditing tools for visualizing privacy leakage.
Problem

Research questions and friction points this paper is trying to address.

Assessing privacy risks in machine learning models using diffusion models.
Evaluating the effectiveness of differential privacy against data reconstruction attacks.
Investigating the impact of real-world data priors on privacy leakage.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses diffusion models for privacy auditing
Targets differential privacy defenses effectively
Visualizes privacy leakage with data priors
🔎 Similar Papers
Kristian Schwethelm
Kristian Schwethelm
Technical University of Munich
Large Language ModelsTrustworthy AIPrivacy-Preserving ML
J
Johannes Kaiser
Chair for Artificial Intelligence in Medicine, Technical University of Munich, Germany
M
Moritz Knolle
Chair for Artificial Intelligence in Medicine, Technical University of Munich, Germany
D
D. Rueckert
Department of Computing, Imperial College London, United Kingdom
G
G. Kaissis
Institute for Machine Learning in Biomedical Imaging, Helmholtz Munich, Germany
Alexander Ziller
Alexander Ziller
Technische Universität München
Privacy-preserving Machine LearningAI in HealthComputer Vision