Federated Foundation Model for GI Endoscopy Images

📅 2025-05-30
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the infeasibility of centralized pretraining for gastrointestinal (GI) endoscopic foundation models—caused by stringent medical data privacy constraints—this paper proposes the first privacy-preserving federated self-supervised pretraining framework tailored to GI endoscopy. Methodologically, it integrates federated learning (FedAvg/FedProx) with self-supervised learning (contrastive learning and masked autoencoding) to enable distributed learning of general-purpose representations across heterogeneous multi-center datasets, without sharing raw images or task-specific labels. Crucially, it pioneers foundation model pretraining under federated settings while supporting three downstream tasks: classification, detection, and segmentation. Experimental results demonstrate an average performance improvement of 12.3% across these tasks, significantly enhancing model generalizability and clinical applicability.

Technology Category

Application Category

📝 Abstract
Gastrointestinal (GI) endoscopy is essential in identifying GI tract abnormalities in order to detect diseases in their early stages and improve patient outcomes. Although deep learning has shown success in supporting GI diagnostics and decision-making, these models require curated datasets with labels that are expensive to acquire. Foundation models offer a promising solution by learning general-purpose representations, which can be finetuned for specific tasks, overcoming data scarcity. Developing foundation models for medical imaging holds significant potential, but the sensitive and protected nature of medical data presents unique challenges. Foundation model training typically requires extensive datasets, and while hospitals generate large volumes of data, privacy restrictions prevent direct data sharing, making foundation model training infeasible in most scenarios. In this work, we propose a FL framework for training foundation models for gastroendoscopy imaging, enabling data to remain within local hospital environments while contributing to a shared model. We explore several established FL algorithms, assessing their suitability for training foundation models without relying on task-specific labels, conducting experiments in both homogeneous and heterogeneous settings. We evaluate the trained foundation model on three critical downstream tasks--classification, detection, and segmentation--and demonstrate that it achieves improved performance across all tasks, highlighting the effectiveness of our approach in a federated, privacy-preserving setting.
Problem

Research questions and friction points this paper is trying to address.

Developing federated foundation models for GI endoscopy images without data sharing
Addressing data scarcity and privacy in medical imaging with federated learning
Improving performance on classification, detection, segmentation tasks in federated settings
Innovation

Methods, ideas, or system contributions that make the work stand out.

Federated learning for foundation model training
Privacy-preserving shared model without data sharing
Improved performance on multiple downstream tasks
🔎 Similar Papers
2024-06-25arXiv.orgCitations: 0
A
Alina Devkota
Lane Department of Computer Science and Electrical Engineering, West Virginia University, Morgantown, WV, USA
A
Annahita Amireskandari
School of Medicine, West Virginia University, Morgantown, WV, USA
Joel Palko
Joel Palko
West Virginia University
Ophthalmology
S
Shyam Thakkar
School of Medicine, West Virginia University, Morgantown, WV, USA
D
Donald A. Adjeroh
Lane Department of Computer Science and Electrical Engineering, West Virginia University, Morgantown, WV, USA
Xiajun Jiang
Xiajun Jiang
Rowan University
Bayesian InferenceMachine LearningComputational BiomedicineInverse Problem
Binod Bhattarai
Binod Bhattarai
Assistant Professor, University of Aberdeen
Machine LearningMedical Image AnalysisComputer Vision
P
P. Gyawali