IndiSeek learns information-guided disentangled representations

๐Ÿ“… 2025-09-25
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
In multimodal single-cell analysis, a key challenge in disentangled representation learning is to jointly learn shared features that are mutually independent and informationally complete, alongside modality-specific features. To address this, we propose IndiSeek: a variational disentanglement framework grounded in the information bottleneck principle. IndiSeek introduces a differentiable reconstruction loss guided by a lower bound on conditional mutual information, enabling joint optimization of feature independence and intra-modality information completeness. By leveraging variational inference to approximate intractable mutual information terms, it avoids adversarial training or hard constraints, ensuring fully end-to-end differentiability. Evaluated on synthetic data, CITE-seq, and multiple real-world multimodal benchmarks, IndiSeek achieves superior feature disentanglement quality and establishes new state-of-the-art performance across downstream tasksโ€”including cell type annotation, batch correction, and cross-modality imputation.

Technology Category

Application Category

๐Ÿ“ Abstract
Learning disentangled representations is a fundamental task in multi-modal learning. In modern applications such as single-cell multi-omics, both shared and modality-specific features are critical for characterizing cell states and supporting downstream analyses. Ideally, modality-specific features should be independent of shared ones while also capturing all complementary information within each modality. This tradeoff is naturally expressed through information-theoretic criteria, but mutual-information-based objectives are difficult to estimate reliably, and their variational surrogates often underperform in practice. In this paper, we introduce IndiSeek, a novel disentangled representation learning approach that addresses this challenge by combining an independence-enforcing objective with a computationally efficient reconstruction loss that bounds conditional mutual information. This formulation explicitly balances independence and completeness, enabling principled extraction of modality-specific features. We demonstrate the effectiveness of IndiSeek on synthetic simulations, a CITE-seq dataset and multiple real-world multi-modal benchmarks.
Problem

Research questions and friction points this paper is trying to address.

Learning disentangled representations for multi-modal data
Separating modality-specific from shared features effectively
Overcoming mutual information estimation challenges in representation learning
Innovation

Methods, ideas, or system contributions that make the work stand out.

IndiSeek combines independence objective with reconstruction loss
It balances independence and completeness in feature extraction
Method enables principled extraction of modality-specific features
๐Ÿ”Ž Similar Papers
No similar papers found.
Yu Gui
Yu Gui
the Wharton School, University of Pennsylvania
Statisticsdistribution-free inferencetransfer learningrepresentation learning
C
Cong Ma
Department of Statistics, University of Chicago
Z
Zongming Ma
Department of Statistics and Data Science, Yale University