š¤ AI Summary
This work identifies the fundamental mechanism underlying the degraded critical capacity of standard binary Restricted Boltzmann Machines (RBMs) when used as associative memories: anomalous phase transitions arising from overly simplistic hidden-unit priors and global connectivity. To address this, we propose an enhanced RBM architecture incorporating local biases and expressive priorsāsuch as multi-state or ReLU-like distributionsāfor hidden units. Leveraging statistical physics and disordered-system theory, we systematically analyze the impact of diverse hidden-unit priors on phase-diagram structure and finite-temperature memory retrieval performance, using Monte Carlo simulations and energy landscape modeling. Results demonstrate that the proposed architecture restores a robust ordered retrieval phase, substantially increasing memory capacity and noise robustness. This study provides thermodynamically grounded theoretical principles and a practical architectural paradigm for designing interpretable latent spaces in generative models.
š Abstract
We investigate the phase diagram and memory retrieval capabilities of bipartite energy-based neural networks, namely Restricted Boltzmann Machines (RBMs), as a function of the prior distribution imposed on their hidden units - including binary, multi-state, and ReLU-like activations. Drawing connections to the Hopfield model and employing analytical tools from statistical physics of disordered systems, we explore how the architectural choices and activation functions shape the thermodynamic properties of these models. Our analysis reveals that standard RBMs with binary hidden nodes and extensive connectivity suffer from reduced critical capacity, limiting their effectiveness as associative memories. To address this, we examine several modifications, such as introducing local biases and adopting richer hidden unit priors. These adjustments restore ordered retrieval phases and markedly improve recall performance, even at finite temperatures. Our theoretical findings, supported by finite-size Monte Carlo simulations, highlight the importance of hidden unit design in enhancing the expressive power of RBMs.