Mix-modal Federated Learning for MRI Image Segmentation

📅 2025-09-02
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
In decentralized medical settings, existing centralized multimodal segmentation methods fail due to severe heterogeneity across hospitals in both MRI modality composition and data distribution—termed *dual heterogeneity* (modality and data). To address this, we propose MixMFL, the first federated learning paradigm for decentralized hybrid-modality MRI segmentation. Its core innovations are: (i) a *modality-decoupled encoder* that disentangles modality-specific and shared latent representations; and (ii) a *dynamic modality prototype memory* enabling adaptive inter-client modality fusion and missing-modality compensation during federated aggregation. Extensive experiments on two public MRI datasets demonstrate that MixMFL significantly outperforms state-of-the-art federated segmentation methods, achieving superior robustness—particularly on clients with incomplete modality sets—while maintaining high segmentation accuracy under realistic cross-hospital data heterogeneity.

Technology Category

Application Category

📝 Abstract
Magnetic resonance imaging (MRI) image segmentation is crucial in diagnosing and treating many diseases, such as brain tumors. Existing MRI image segmentation methods mainly fall into a centralized multimodal paradigm, which is inapplicable in engineering non-centralized mix-modal medical scenarios. In this situation, each distributed client (hospital) processes multiple mixed MRI modalities, and the modality set and image data for each client are diverse, suffering from extensive client-wise modality heterogeneity and data heterogeneity. In this paper, we first formulate non-centralized mix-modal MRI image segmentation as a new paradigm for federated learning (FL) that involves multiple modalities, called mix-modal federated learning (MixMFL). It distinguishes from existing multimodal federating learning (MulMFL) and cross-modal federating learning (CroMFL) paradigms. Then, we proposed a novel modality decoupling and memorizing mix-modal federated learning framework (MDM-MixMFL) for MRI image segmentation, which is characterized by a modality decoupling strategy and a modality memorizing mechanism. Specifically, the modality decoupling strategy disentangles each modality into modality-tailored and modality-shared information. During mix-modal federated updating, corresponding modality encoders undergo tailored and shared updating, respectively. It facilitates stable and adaptive federating aggregation of heterogeneous data and modalities from distributed clients. Besides, the modality memorizing mechanism stores client-shared modality prototypes dynamically refreshed from every modality-tailored encoder to compensate for incomplete modalities in each local client. It further benefits modality aggregation and fusion processes during mixmodal federated learning. Extensive experiments on two public datasets for MRI image segmentation demonstrate the effectiveness and superiority of our methods.
Problem

Research questions and friction points this paper is trying to address.

Addresses MRI segmentation in decentralized mix-modal medical settings
Handles client-wise modality and data heterogeneity in federated learning
Proposes modality decoupling and memorizing for incomplete modality compensation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Modality decoupling strategy for tailored and shared information
Modality memorizing mechanism storing client-shared prototypes
Dynamic prototype refreshment compensating incomplete local modalities
🔎 Similar Papers
No similar papers found.
G
Guyue Hu
State Key Laboratory of Opto-Electronic Information Acquisition and Protection Technology, Anhui University, Hefei, 230601, China; Key Laboratory of Intelligent Computing and Signal Processing of Ministry of Education, Anhui University, Hefei, 230601, China; Anhui Provincial Key Laboratory of Security Artificial Intelligence, Anhui University, Hefei, 230601, China; Anhui Provincial Key Laboratory of Multimodal Cognitive Computation, Anhui University, 230601, Hefei, China; School of Artificial Intelligence,
Siyuan Song
Siyuan Song
Associate Professor, Arizona State University
Construction Safety and HealthWorkforce DevelopmentAI in ConstructionEngineering Education
Jingpeng Sun
Jingpeng Sun
State Key Laboratory of Opto-Electronic Information Acquisition and Protection Technology, Anhui University, Hefei, 230601, China; Key Laboratory of Intelligent Computing and Signal Processing of Ministry of Education, Anhui University, Hefei, 230601, China; Anhui Provincial Key Laboratory of Security Artificial Intelligence, Anhui University, Hefei, 230601, China; Anhui Provincial Key Laboratory of Multimodal Cognitive Computation, Anhui University, 230601, Hefei, China; School of Artificial Intelligence,
Z
Zhe Jin
State Key Laboratory of Opto-Electronic Information Acquisition and Protection Technology, Anhui University, Hefei, 230601, China; Anhui Provincial Key Laboratory of Security Artificial Intelligence, Anhui University, Hefei, 230601, China; School of Artificial Intelligence, Anhui University, 230601, Hefei, China
Chenglong Li
Chenglong Li
Professor, The University of Florida
Drug DesignDrug DiscoveryMolecular RecognitionMolecular ModelingProtein structure and Dynamics
Jin Tang
Jin Tang
Anhui University
Computer visionintelligent video analysis