Enhancing diffusion models with Gaussianization preprocessing

📅 2025-12-24
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Diffusion models suffer from slow sampling and poor early-stage reconstruction quality—particularly in lightweight architectures—due to trajectory branching delays. To address this, we propose Gaussianization preprocessing: prior to training, image data are transformed into an approximately independent standard Gaussian distribution via invertible mappings (e.g., normalizing flows or empirical CDF transforms). This is the first integration of such reversible Gaussianization into the diffusion training pipeline, aiming to mitigate branching at the data-distribution level. Experiments across multiple image benchmarks demonstrate that, under small UNet architectures, our method improves PSNR by 12% in early denoising steps, reduces sampling steps by 30% without sacrificing visual fidelity, and yields significant FID reduction alongside accelerated convergence. Our core contribution lies in reshaping the data distribution to lower task complexity, thereby enhancing stability, efficiency, and early reconstruction quality—especially for resource-constrained diffusion models.

Technology Category

Application Category

📝 Abstract
Diffusion models are a class of generative models that have demonstrated remarkable success in tasks such as image generation. However, one of the bottlenecks of these models is slow sampling due to the delay before the onset of trajectory bifurcation, at which point substantial reconstruction begins. This issue degrades generation quality, especially in the early stages. Our primary objective is to mitigate bifurcation-related issues by preprocessing the training data to enhance reconstruction quality, particularly for small-scale network architectures. Specifically, we propose applying Gaussianization preprocessing to the training data to make the target distribution more closely resemble an independent Gaussian distribution, which serves as the initial density of the reconstruction process. This preprocessing step simplifies the model's task of learning the target distribution, thereby improving generation quality even in the early stages of reconstruction with small networks. The proposed method is, in principle, applicable to a broad range of generative tasks, enabling more stable and efficient sampling processes.
Problem

Research questions and friction points this paper is trying to address.

Improves early-stage image generation quality in diffusion models
Accelerates sampling by preprocessing data to resemble Gaussian distribution
Enables stable sampling with small-scale network architectures
Innovation

Methods, ideas, or system contributions that make the work stand out.

Gaussianization preprocessing for training data
Making target distribution resemble independent Gaussian
Enhancing early-stage reconstruction in small networks
🔎 Similar Papers
No similar papers found.
L
Li Cunzhi
Graduate School of Informatics, Kyoto University, Kyoto, Japan
L
Louis Kang
Graduate School of Informatics, Kyoto University, Kyoto, Japan; Neural Circuits and Computations Unit, RIKEN Center for Brain Science, Saitama, Japan
Hideaki Shimazaki
Hideaki Shimazaki
Kyoto University
Theoretical Neuroscience