Particle Dynamics for Latent-Variable Energy-Based Models

๐Ÿ“… 2025-10-17
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
This paper addresses the maximum-likelihood training challenge for latent-variable energy-based models (LVEBMs). We propose a manifold-based saddle-point optimization framework grounded in particle dynamics. Specifically, training is reformulated as a saddle-point problem over a latent manifold, where the joint distribution and latent variable distribution are jointly optimized. Coupled Wasserstein gradient flows alternately update joint negative samples and latent particles, numerically simulated via overdamped Langevin dynamics and stochastic parameter ascentโ€”without requiring discriminators or auxiliary networks. Theoretically, our derived evidence lower bound (ELBO) strictly improves upon standard variational inference and enjoys guaranteed convergence. Empirically, the method achieves high-fidelity generation and strong disentanglement of latent variables on tasks such as physical system modeling, matching or surpassing state-of-the-art approaches in performance.

Technology Category

Application Category

๐Ÿ“ Abstract
Latent-variable energy-based models (LVEBMs) assign a single normalized energy to joint pairs of observed data and latent variables, offering expressive generative modeling while capturing hidden structure. We recast maximum-likelihood training as a saddle problem over distributions on the latent and joint manifolds and view the inner updates as coupled Wasserstein gradient flows. The resulting algorithm alternates overdamped Langevin updates for a joint negative pool and for conditional latent particles with stochastic parameter ascent, requiring no discriminator or auxiliary networks. We prove existence and convergence under standard smoothness and dissipativity assumptions, with decay rates in KL divergence and Wasserstein-2 distance. The saddle-point view further yields an ELBO strictly tighter than bounds obtained with restricted amortized posteriors. Our method is evaluated on numerical approximations of physical systems and performs competitively against comparable approaches.
Problem

Research questions and friction points this paper is trying to address.

Training latent-variable energy-based models via saddle-point optimization
Developing discriminator-free training using coupled Wasserstein gradient flows
Providing theoretical convergence guarantees for generative modeling performance
Innovation

Methods, ideas, or system contributions that make the work stand out.

Latent-variable energy-based models with joint energy
Maximum-likelihood training as saddle problem
Alternating Langevin updates without auxiliary networks
๐Ÿ”Ž Similar Papers
No similar papers found.
Shiqin Tang
Shiqin Tang
Center for AI and Robotics, Chinese Academy of Sciences
Machine Learning
S
Shuxin Zhuang
Department of Data Science, City University of Hong Kong, Hong Kong
R
Rong Feng
Department of Computer Science, City University of Hong Kong, Hong Kong
Runsheng Yu
Runsheng Yu
Unknown affiliation
H
Hongzong Li
Generative AI Research and Development Center, The Hong Kong University of Science and Technology, Hong Kong
Youzhi Zhang
Youzhi Zhang
CAIR, Hong Kong Institute of Science & Innovation, Chinese Academy of Sciences
Computational Game TheoryOptimizationMulti-agent SystemsArtificial Intelligence