🤖 AI Summary
This work investigates the generalization and memorization behaviors of denoising score matching (DSM) in diffusion models under high-dimensional, large-data, and multi-feature regimes. We analyze random feature networks approximating the score function under standard Gaussian data, and—crucially—derive exact analytical expressions for both training and test errors of DSM in the high-dimensional asymptotic limit. Leveraging tools from random matrix theory and statistical physics, we quantitatively characterize how model complexity $p$ and the number $m$ of noisy samples govern the generalization–memorization phase transition, precisely identifying the critical phase boundary. Our theoretical predictions exhibit excellent agreement with numerical experiments, uncovering the critical mechanism underlying the transition from generalization to memorization in the overparameterized regime. The results provide an interpretable theoretical foundation for diffusion model architecture design, training strategy optimization, and controllable generalization performance.
📝 Abstract
We derive asymptotically precise expressions for test and train errors of denoising score matching (DSM) in generative diffusion models. The score function is parameterized by random features neural networks, with the target distribution being $d$-dimensional standard Gaussian. We operate in a regime where the dimension $d$, number of data samples $n$, and number of features $p$ tend to infinity while keeping the ratios $psi_n=frac{n}{d}$ and $psi_p=frac{p}{d}$ fixed. By characterizing the test and train errors, we identify regimes of generalization and memorization in diffusion models. Furthermore, our work sheds light on the conditions enhancing either generalization or memorization. Consistent with prior empirical observations, our findings indicate that the model complexity ($p$) and the number of noise samples per data sample ($m$) used during DSM significantly influence generalization and memorization behaviors.