Test-time Adaptation for Foundation Medical Segmentation Model without Parametric Updates

📅 2025-04-02
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Foundational medical segmentation models (e.g., MedSAM) exhibit limited performance on complex lesions and are highly sensitive to bounding-box prompt perturbations. Existing test-time adaptation (TTA) methods suffer from weak update signals, catastrophic forgetting, and prohibitive computational overhead. To address these issues, we propose a **parameter-free TTA paradigm**: we theoretically establish that optimizing image embeddings is equivalent to fine-tuning model parameters; further, we design a joint objective combining distribution-approximated implicit conditional random field loss and entropy minimization to achieve robust embedding-space adaptation. This approach significantly mitigates forgetting and reduces computational complexity. Evaluated on three medical segmentation benchmarks, our method achieves an average Dice score improvement of ~3% and reduces inference overhead by over 7×, striking an effective balance between efficiency and accuracy.

Technology Category

Application Category

📝 Abstract
Foundation medical segmentation models, with MedSAM being the most popular, have achieved promising performance across organs and lesions. However, MedSAM still suffers from compromised performance on specific lesions with intricate structures and appearance, as well as bounding box prompt-induced perturbations. Although current test-time adaptation (TTA) methods for medical image segmentation may tackle this issue, partial (e.g., batch normalization) or whole parametric updates restrict their effectiveness due to limited update signals or catastrophic forgetting in large models. Meanwhile, these approaches ignore the computational complexity during adaptation, which is particularly significant for modern foundation models. To this end, our theoretical analyses reveal that directly refining image embeddings is feasible to approach the same goal as parametric updates under the MedSAM architecture, which enables us to realize high computational efficiency and segmentation performance without the risk of catastrophic forgetting. Under this framework, we propose to encourage maximizing factorized conditional probabilities of the posterior prediction probability using a proposed distribution-approximated latent conditional random field loss combined with an entropy minimization loss. Experiments show that we achieve about 3% Dice score improvements across three datasets while reducing computational complexity by over 7 times.
Problem

Research questions and friction points this paper is trying to address.

Improving segmentation of complex lesions without parametric updates
Reducing computational complexity in foundation model adaptation
Avoiding catastrophic forgetting during test-time adaptation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Refining image embeddings without parametric updates
Using distribution-approximated latent conditional random field loss
Combining entropy minimization loss for improved performance
🔎 Similar Papers
No similar papers found.
Kecheng Chen
Kecheng Chen
PhD student at EE, City University of Hong Kong
Transfer LearningAI for HealthcareSignal Processing
X
Xinyu Luo
Department of Electrical Engineering and the CIMDA, City University of Hong Kong
Tiexin Qin
Tiexin Qin
City University of Hong Kong
Machine LearningDynamical Systems
J
Jie Liu
Department of Electrical Engineering and the CIMDA, City University of Hong Kong
H
Hui Liu
Department of Electrical Engineering and the CIMDA, City University of Hong Kong
V
Victor Ho-Fun Lee
Department of Clinical Oncology, The University of Hong Kong
H
Hong Yan
Department of Electrical Engineering and the CIMDA, City University of Hong Kong
Haoliang Li
Haoliang Li
Department of Electrical Engineering, City University of Hong Kong
AI SecurityInformation Forensics and SecurityMachine Learning