🤖 AI Summary
Clinical ophthalmic multimodal data often suffer from missing modalities, high redundancy, and severe inter-modal representation coupling. To address these challenges, we propose an end-to-end Essence-Point extraction and Disentangled Representation Learning framework (EDRL). Our key contributions are: (i) the novel Essence-Point representation learning module, which employs contrastive feature selection to precisely extract discriminative local essence points while suppressing irrelevant information (e.g., redundant slices); and (ii) a dual-branch disentanglement mechanism—separating modality-shared and modality-specific representations—integrated with self-distillation and disentangled fusion for explicit cross-modal representation separation. Evaluated on multiple real-world ophthalmic multimodal datasets, EDRL achieves an average 4.2% improvement in grading accuracy over state-of-the-art methods, enhances robustness by 37% under single-modality missing scenarios, and significantly improves model interpretability and clinical applicability.
📝 Abstract
This paper discusses how ophthalmologists often rely on multimodal data to improve diagnostic accuracy. However, complete multimodal data is rare in real-world applications due to a lack of medical equipment and concerns about data privacy. Traditional deep learning methods typically address these issues by learning representations in latent space. However, the paper highlights two key limitations of these approaches: (i) Task-irrelevant redundant information (e.g., numerous slices) in complex modalities leads to significant redundancy in latent space representations. (ii) Overlapping multimodal representations make it difficult to extract unique features for each modality. To overcome these challenges, the authors propose the Essence-Point and Disentangle Representation Learning (EDRL) strategy, which integrates a self-distillation mechanism into an end-to-end framework to enhance feature selection and disentanglement for more robust multimodal learning. Specifically, the Essence-Point Representation Learning module selects discriminative features that improve disease grading performance. The Disentangled Representation Learning module separates multimodal data into modality-common and modality-unique representations, reducing feature entanglement and enhancing both robustness and interpretability in ophthalmic disease diagnosis. Experiments on multimodal ophthalmology datasets show that the proposed EDRL strategy significantly outperforms current state-of-the-art methods.