BLOSSOM: Block-wise Federated Learning Over Shared and Sparse Observed Modalities

📅 2026-03-29
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the performance degradation in real-world multimodal federated learning caused by missing and heterogeneous modalities across clients. To tackle this challenge, the authors propose a task-agnostic, block-level federated learning framework that employs a modular neural network architecture. This design enables flexible participation from clients with arbitrary modality subsets through a block-wise parameter aggregation mechanism. Furthermore, it integrates modality-aware personalized training to preserve task-specific representations while sharing common modules across participants. Experimental results demonstrate that the proposed framework achieves an average performance gain of 18.7% in scenarios with incomplete modalities and up to 37.7% improvement when clients possess exclusive modalities, substantially enhancing the practicality and robustness of multimodal federated learning systems.
📝 Abstract
Multimodal federated learning (FL) is essential for real-world applications such as autonomous systems and healthcare, where data is distributed across heterogeneous clients with varying and often missing modalities. However, most existing FL approaches assume uniform modality availability, limiting their applicability in practice. We introduce BLOSSOM, a task-agnostic framework for multimodal FL designed to operate under shared and sparsely observed modality conditions. BLOSSOM supports clients with arbitrary modality subsets and enables flexible sharing of model components. To address client and task heterogeneity, we propose a block-wise aggregation strategy that selectively aggregates shared components while keeping task-specific blocks private, enabling partial personalization. We evaluate BLOSSOM on multiple diverse multimodal datasets and analyse the effects of missing modalities and personalization. Our results show that block-wise personalization significantly improves performance, particularly in settings with severe modality sparsity. In modality-incomplete scenarios, BLOSSOM achieves an average performance gain of 18.7% over full-model aggregation, while in modality-exclusive settings the gain increases to 37.7%, highlighting the importance of block-wise learning for practical multimodal FL systems.
Problem

Research questions and friction points this paper is trying to address.

multimodal federated learning
modality sparsity
client heterogeneity
missing modalities
federated learning
Innovation

Methods, ideas, or system contributions that make the work stand out.

block-wise aggregation
multimodal federated learning
modality sparsity
partial personalization
heterogeneous clients
🔎 Similar Papers
No similar papers found.
P
Pranav M R
DaSH Lab, BITS Pilani, KK Birla Goa Campus, India
J
Jayant Chandwani
DaSH Lab, BITS Pilani, KK Birla Goa Campus, India
A
Ahmed M. Abdelmoniem
Queen Mary University of London, United Kingdom
Arnab K. Paul
Arnab K. Paul
Assistant Professor @ CSIS in BITS Pilani - Goa Campus, PostDoc - ORNL, PhD - Virginia Tech
Distributed SystemsStorage SystemsHigh Performance ComputingInternet-of-ThingsCloud Computing