ReTrack: Data Unlearning in Diffusion Models through Redirecting the Denoising Trajectory

📅 2025-09-16
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Diffusion models risk memorizing training data, compromising user privacy and model security. To address this, we propose an efficient data unlearning method specifically designed for diffusion models—without retraining. Our approach employs importance sampling to construct an interpretable fine-tuning loss that selectively emphasizes dominant denoising terms. Crucially, we redirect the target denoising trajectory toward the *k*-nearest neighbors of the to-be-forgotten sample, enabling semantically consistent unlearning guidance. By jointly optimizing an approximate loss and steering the denoising trajectory, our method achieves superior unlearning performance on MNIST T-Shirt, CelebA-HQ, CIFAR-10, and Stable Diffusion—outperforming existing approaches across diverse benchmarks. Notably, it maintains high generation fidelity even under strong unlearning intensity, marking the first work to achieve an optimal trade-off between unlearning effectiveness and generative fidelity in diffusion models.

Technology Category

Application Category

📝 Abstract
Diffusion models excel at generating high-quality, diverse images but suffer from training data memorization, raising critical privacy and safety concerns. Data unlearning has emerged to mitigate this issue by removing the influence of specific data without retraining from scratch. We propose ReTrack, a fast and effective data unlearning method for diffusion models. ReTrack employs importance sampling to construct a more efficient fine-tuning loss, which we approximate by retaining only dominant terms. This yields an interpretable objective that redirects denoising trajectories toward the $k$-nearest neighbors, enabling efficient unlearning while preserving generative quality. Experiments on MNIST T-Shirt, CelebA-HQ, CIFAR-10, and Stable Diffusion show that ReTrack achieves state-of-the-art performance, striking the best trade-off between unlearning strength and generation quality preservation.
Problem

Research questions and friction points this paper is trying to address.

Mitigate training data memorization in diffusion models
Remove specific data influence without full retraining
Redirect denoising trajectories for efficient unlearning
Innovation

Methods, ideas, or system contributions that make the work stand out.

Redirects denoising trajectories
Uses importance sampling loss
Approximates dominant terms efficiently
🔎 Similar Papers
No similar papers found.
Q
Qitan Shi
Department of Electronic Engineering, Tsinghua University
C
Cheng Jin
Department of Electronic Engineering, Tsinghua University
J
Jiawei Zhang
Department of Electronic Engineering, Tsinghua University
Yuantao Gu
Yuantao Gu
Department of Electronic Engineering, Tsinghua University
Signal processingSparse recoverySparse learningOptimizationGraph signal processing