Approximating Two-Layer ReLU Networks for Hidden State Analysis in Differential Privacy

📅 2024-07-05
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing privacy analyses under the hidden-state threat model with differential privacy are restricted to convex optimization, failing to provide rigorous privacy guarantees for multilayer non-convex neural networks. This work is the first to extend this model to training single-hidden-layer ReLU networks. We propose a stochastic dual approximation method to construct a strongly convex surrogate problem and design the Noisy Cyclic Gradient Descent (NoisyCGD) algorithm. Theoretically, we establish convergence acceleration and provable differential privacy under non-convex settings. Empirically, on standard classification benchmarks, NoisyCGD achieves utility comparable to or better than DP-SGD under identical privacy budgets, with significantly improved accuracy. Our approach breaks the convexity dependency of the hidden-state model, delivering the first rigorous differentially private training framework applicable to non-convex deep learning architectures.

Technology Category

Application Category

📝 Abstract
The hidden state threat model of differential privacy (DP) assumes that the adversary has access only to the final trained machine learning (ML) model, without seeing intermediate states during training. Current privacy analyses under this model, however, are limited to convex optimization problems, reducing their applicability to multi-layer neural networks, which are essential in modern deep learning applications. Additionally, the most successful applications of the hidden state privacy analyses in classification tasks have been for logistic regression models. We demonstrate that it is possible to privately train convex problems with privacy-utility trade-offs comparable to those of one hidden-layer ReLU networks trained with DP stochastic gradient descent (DP-SGD). We achieve this through a stochastic approximation of a dual formulation of the ReLU minimization problem which results in a strongly convex problem. This enables the use of existing hidden state privacy analyses, providing accurate privacy bounds also for the noisy cyclic mini-batch gradient descent (NoisyCGD) method with fixed disjoint mini-batches. Our experiments on benchmark classification tasks show that NoisyCGD can achieve privacy-utility trade-offs comparable to DP-SGD applied to one-hidden-layer ReLU networks. Additionally, we provide theoretical utility bounds that highlight the speed-ups gained through the convex approximation.
Problem

Research questions and friction points this paper is trying to address.

Extends hidden state DP analysis to non-convex neural networks
Enables private training comparable to DP-SGD for ReLU networks
Provides accurate privacy bounds for NoisyCGD method
Innovation

Methods, ideas, or system contributions that make the work stand out.

Convex approximation of ReLU networks
Stochastic dual formulation for privacy
NoisyCGD matches DP-SGD performance