Canonical normalizing flows for manifold learning

📅 2023-10-19
🏛️ Neural Information Processing Systems
📈 Citations: 7
Influential: 0
📄 PDF
🤖 AI Summary
To address dimension redundancy, feature entanglement, and basis vector degeneracy in manifold learning, this paper proposes an invertible manifold modeling framework grounded in a *canonical intrinsic basis*. The method jointly learns a sparse, near-orthogonal, and non-degenerate latent-space eigenbasis—defined as the canonical intrinsic basis—by imposing ℓ₁-norm regularization on the manifold metric tensor to enable automatic basis selection. It integrates canonical normalization flows with manifold-constrained invertible mappings to preserve geometric consistency. Optimized within a maximum-likelihood framework, the approach significantly improves latent-space utilization and density estimation accuracy. Experiments demonstrate a 30–50% reduction in effective dimensionality and an average 12.6% improvement in Fréchet Inception Distance (FID) over state-of-the-art manifold-based normalizing flows, establishing superior performance across multiple benchmarks.
📝 Abstract
Manifold learning flows are a class of generative modelling techniques that assume a low-dimensional manifold description of the data. The embedding of such a manifold into the high-dimensional space of the data is achieved via learnable invertible transformations. Therefore, once the manifold is properly aligned via a reconstruction loss, the probability density is tractable on the manifold and maximum likelihood can be used to optimize the network parameters. Naturally, the lower-dimensional representation of the data requires an injective-mapping. Recent approaches were able to enforce that the density aligns with the modelled manifold, while efficiently calculating the density volume-change term when embedding to the higher-dimensional space. However, unless the injective-mapping is analytically predefined, the learned manifold is not necessarily an efficient representation of the data. Namely, the latent dimensions of such models frequently learn an entangled intrinsic basis, with degenerate information being stored in each dimension. Alternatively, if a locally orthogonal and/or sparse basis is to be learned, here coined canonical intrinsic basis, it can serve in learning a more compact latent space representation. Toward this end, we propose a canonical manifold learning flow method, where a novel optimization objective enforces the transformation matrix to have few prominent and non-degenerate basis functions. We demonstrate that by minimizing the off-diagonal manifold metric elements $ell_1$-norm, we can achieve such a basis, which is simultaneously sparse and/or orthogonal. Canonical manifold flow yields a more efficient use of the latent space, automatically generating fewer prominent and distinct dimensions to represent data, and a better approximation of target distributions than other manifold flow methods in most experiments we conducted, resulting in lower FID scores.
Problem

Research questions and friction points this paper is trying to address.

Develops canonical manifold learning flows for efficient data representation.
Enforces sparse and orthogonal basis in latent space optimization.
Improves target distribution approximation with lower FID scores.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Canonical manifold learning flow method
Optimization enforces sparse orthogonal basis
Minimizes off-diagonal manifold metric elements
🔎 Similar Papers