๐ค AI Summary
Traditional DP-SGD employs fixed-form noise (e.g., Gaussian or Laplace), where a single parameter governs both privacy loss and utility degradation, resulting in strong coupling that impedes flexible trade-offs across diverse training configurations (e.g., varying iteration count $T$ or batch size $B$). To address this, we propose PLRV-O, a framework that breaks the single-degree-of-freedom constraint by directly optimizing the Privacy Loss Random Variable (PLRV) to construct a tunable noise distribution search spaceโenabling decoupled optimization of privacy and utility losses. PLRV-O supports adaptive configuration across model scale, training duration, and sampling strategy, and integrates a tight moment accountant for rigorous privacy accounting. Experiments demonstrate substantial improvements: on CIFAR-10 at $varepsilon approx 0.5$, accuracy reaches 94.03% (+10.1 percentage points); on SST-2 at $varepsilon approx 0.2$, it achieves 92.20% (+41.95 percentage points), significantly outperforming standard DP-SGD baselines.
๐ Abstract
Differentially Private Stochastic Gradient Descent (DP-SGD) is a standard method for enforcing privacy in deep learning, typically using the Gaussian mechanism to perturb gradient updates. However, conventional mechanisms such as Gaussian and Laplacian noise are parameterized only by variance or scale. This single degree of freedom ties the magnitude of noise directly to both privacy loss and utility degradation, preventing independent control of these two factors. The problem becomes more pronounced when the number of composition rounds T and batch size B vary across tasks, as these variations induce task-dependent shifts in the privacy-utility trade-off, where small changes in noise parameters can disproportionately affect model accuracy. To address this limitation, we introduce PLRV-O, a framework that defines a broad search space of parameterized DP-SGD noise distributions, where privacy loss moments are tightly characterized yet can be optimized more independently with respect to utility loss. This formulation enables systematic adaptation of noise to task-specific requirements, including (i) model size, (ii) training duration, (iii) batch sampling strategies, and (iv) clipping thresholds under both training and fine-tuning settings. Empirical results demonstrate that PLRV-O substantially improves utility under strict privacy constraints. On CIFAR-10, a fine-tuned ViT achieves 94.03% accuracy at epsilon approximately 0.5, compared to 83.93% with Gaussian noise. On SST-2, RoBERTa-large reaches 92.20% accuracy at epsilon approximately 0.2, versus 50.25% with Gaussian.