Bounding Reconstruction Attack Success of Adversaries Without Data Priors

📅 2024-02-20
🏛️ arXiv.org
📈 Citations: 1
Influential: 0
📄 PDF
🤖 AI Summary
In differential privacy (DP), the privacy budget ε lacks a quantitative mapping to the risk of training data reconstruction attacks, especially under realistic assumptions—i.e., without data priors and beyond worst-case scenarios. Method: We propose the first verifiable, formal upper bound on training data reconstruction success rate, integrating rigorous DP theoretical analysis, information-theoretic bounding techniques, and gradient-leakage modeling. The bound is empirically validated via inversion attacks and other reconstruction benchmarks. Contribution/Results: Our bound is tight, supports multi-metric and multi-scenario ε selection, and—crucially—establishes the first quantitative relationship between ε and reconstruction success probability. It enables principled, scientifically grounded DP hyperparameter tuning, significantly improving the efficiency and rigor of the privacy–utility trade-off.

Technology Category

Application Category

📝 Abstract
Reconstruction attacks on machine learning (ML) models pose a strong risk of leakage of sensitive data. In specific contexts, an adversary can (almost) perfectly reconstruct training data samples from a trained model using the model's gradients. When training ML models with differential privacy (DP), formal upper bounds on the success of such reconstruction attacks can be provided. So far, these bounds have been formulated under worst-case assumptions that might not hold high realistic practicality. In this work, we provide formal upper bounds on reconstruction success under realistic adversarial settings against ML models trained with DP and support these bounds with empirical results. With this, we show that in realistic scenarios, (a) the expected reconstruction success can be bounded appropriately in different contexts and by different metrics, which (b) allows for a more educated choice of a privacy parameter.
Problem

Research questions and friction points this paper is trying to address.

Deriving formal privacy bounds for data reconstruction attacks
Bridging the gap between theoretical DP guarantees and real-world threats
Establishing probabilistic success bounds for from-scratch gradient inversion attacks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Derives formal privacy bounds for gradient inversion attacks
Models from-scratch attacks as mean estimation problem
Provides probabilistic success bounds using MSE and PSNR
🔎 Similar Papers
No similar papers found.
Alexander Ziller
Alexander Ziller
Technische Universität München
Privacy-preserving Machine LearningAI in HealthComputer Vision
A
Anneliese Riess
Institute of Machine Learning in Biomedical Imaging, Helmholtz Munich, Neuherberg, Germany
Kristian Schwethelm
Kristian Schwethelm
Technical University of Munich
Large Language ModelsTrustworthy AIPrivacy-Preserving ML
T
Tamara T. Mueller
Chair for Artificial Intelligence in Medicine, Technical University of Munich, Munich, Germany
D
D. Rueckert
Chair for Artificial Intelligence in Medicine, Technical University of Munich, Munich, Germany; Department of Computing, Imperial College London, London, United Kingdom
G
G. Kaissis
Chair for Artificial Intelligence in Medicine, Technical University of Munich, Munich, Germany; Institute of Machine Learning in Biomedical Imaging, Helmholtz Munich, Neuherberg, Germany; Department of Computing, Imperial College London, London, United Kingdom