Safe machine learning model release from Trusted Research Environments: The SACRO-ML package

📅 2022-12-02
📈 Citations: 1
✨ Influential: 0
📄 PDF
🤖 AI Summary
To mitigate statistical disclosure risks associated with releasing machine learning models in trusted research environments, this paper proposes a two-stage (ante-hoc and post-hoc) Statistical Disclosure Control (SDC) framework and open-sources SACRO-ML—a MIT-licensed toolkit. The framework addresses the risk that model parameters may inadvertently reveal statistics about sensitive training data. Its core innovations are: (1) SafeModel, an ante-hoc module that assesses disclosure risk at the model level prior to release; and (2) Attacks, a post-hoc module integrating membership inference, model inversion, and attribute inference attacks to empirically quantify residual risk after training. SACRO-ML supports mainstream ML models and employs scalable encapsulation techniques, significantly enhancing privacy compliance and risk controllability for sensitive-data-driven model deployment. It constitutes the first open-source, system-level implementation of SDC tailored for secure ML model release.
📝 Abstract
We present SACRO-ML, an integrated suite of open source Python tools to facilitate the statistical disclosure control (SDC) of machine learning (ML) models trained on confidential data prior to public release. SACRO-ML combines (i) a SafeModel package that extends commonly used ML models to provide ante-hoc SDC by assessing the vulnerability of disclosure posed by the training regime; and (ii) an Attacks package that provides post-hoc SDC by rigorously assessing the empirical disclosure risk of a model through a variety of simulated attacks after training. The SACRO-ML code and documentation are available under an MIT license at https://github.com/AI-SDC/SACRO-ML
Problem

Research questions and friction points this paper is trying to address.

Ensuring safe release of ML models from confidential data
Providing ante-hoc statistical disclosure control for ML models
Assessing post-hoc disclosure risks through simulated attacks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Open source Python tools for model SDC
SafeModel assesses ante-hoc disclosure vulnerability
Attacks package evaluates post-hoc disclosure risk
🔎 Similar Papers
No similar papers found.
Jim Smith
Jim Smith
Professor in Interactive Artificial Intelligence, University of the West of England
Artificial Intelligence
R
R. Preen
School of Computer Science, University of the West of England
A
Andrew McCarthy
School of Computer Science, University of the West of England
M
Maha Albashir
School of Computer Science, University of the West of England
A
Alba Crespi-Boixader
School of Computer Science, University of the West of England
S
Shahzad Mumtaz
School of Medicine, University of Dundee
J
J. Liley
Department of Mathematical Sciences, University of Durham
S
Simon Rogers
School of Computer Science, University of the West of England
Y
Yola Jones
NHS National Services Scotland