Cascade of phase transitions in the training of Energy-based models

📅 2024-05-23
đŸ›ïž Neural Information Processing Systems
📈 Citations: 2
✹ Influential: 0
📄 PDF
đŸ€– AI Summary
This work investigates the dynamic evolution of feature encoding during Restricted Boltzmann Machine (RBM) training. Methodologically, it combines singular value decomposition to track weight matrix evolution, analytical mean-field analysis, numerical training experiments, and finite-size scaling theory. The key contribution is the first identification—within energy-based generative models (EBMs)—of a mean-field–like paramagnetic–ferromagnetic cascade phase transition sequence during training. Empirically, the transitions sharpen in the high-dimensional limit; the first-order transition belongs to the mean-field universality class; and learning proceeds sequentially: the model first captures the centroid of the empirical data distribution, then progressively extracts principal components. This study establishes a rigorous correspondence between high-dimensional learning dynamics and statistical-physics phase-transition theory, offering a novel paradigm for understanding the intrinsic learning mechanisms of deep generative models.

Technology Category

Application Category

📝 Abstract
In this paper, we investigate the feature encoding process in a prototypical energy-based generative model, the Restricted Boltzmann Machine (RBM). We start with an analytical investigation using simplified architectures and data structures, and end with numerical analysis of real trainings on real datasets. Our study tracks the evolution of the model's weight matrix through its singular value decomposition, revealing a series of phase transitions associated to a progressive learning of the principal modes of the empirical probability distribution. The model first learns the center of mass of the modes and then progressively resolve all modes through a cascade of phase transitions. We first describe this process analytically in a controlled setup that allows us to study analytically the training dynamics. We then validate our theoretical results by training the Bernoulli-Bernoulli RBM on real data sets. By using data sets of increasing dimension, we show that learning indeed leads to sharp phase transitions in the high-dimensional limit. Moreover, we propose and test a mean-field finite-size scaling hypothesis. This shows that the first phase transition is in the same universality class of the one we studied analytically, and which is reminiscent of the mean-field paramagnetic-to-ferromagnetic phase transition.
Problem

Research questions and friction points this paper is trying to address.

Investigates phase transitions in RBM training
Tracks weight matrix evolution via SVD
Validates theory with real dataset experiments
Innovation

Methods, ideas, or system contributions that make the work stand out.

Energy-based generative models
Singular value decomposition
Mean-field scaling hypothesis
🔎 Similar Papers
D
Dimitrios Bachtis
Laboratoire de Physique de l’Ecole Normale SupĂ©rieure, ENS, UniversitĂ© PSL, CNRS, Sorbonne UniversitĂ©, UniversitĂ© Paris CitĂ©, F-75005 Paris, France.
Giulio Biroli
Giulio Biroli
Professor of Theoretical Physics, ENS Paris
Statistical PhysicsCondensed MatterComplex Systems
A
A. Decelle
Departamento de Física Teórica, Universidad Complutense, 28040 Madrid, Spain; Université Paris-Saclay, CNRS, INRIA Tau team, LISN, 91190, Gif-sur-Yvette, France.
B
Beatriz Seoane
Université Paris-Saclay, CNRS, INRIA Tau team, LISN, 91190, Gif-sur-Yvette, France.