🤖 AI Summary
Colonoscopy image analysis faces challenges including pervasive background clutter, complex medical terminology, and ambiguous multi-lesion descriptions—hindering clinical applicability of vision-language joint modeling. To address these, we propose CLEAN (Cleansing-Attunement-Unification), a three-stage progressive self-supervised pretraining framework. Stage one performs frame-level background filtering and LLM-driven clinical attribute extraction to cleanse non-informative images and texts. Stage two introduces patient-level cross-modal attention to disambiguate multiple polyps. Stage three achieves modality unification via fine-grained vision-language contrastive learning. CLEAN integrates CLIP architecture, LLM-based semantic parsing, and self-supervised learning. On zero-shot and few-shot polyp detection and classification tasks, it significantly outperforms state-of-the-art methods, enhancing diagnostic robustness and clinical relevance.
📝 Abstract
Pre-training on image-text colonoscopy records offers substantial potential for improving endoscopic image analysis, but faces challenges including non-informative background images, complex medical terminology, and ambiguous multi-lesion descriptions. We introduce Endo-CLIP, a novel self-supervised framework that enhances Contrastive Language-Image Pre-training (CLIP) for this domain. Endo-CLIP's three-stage framework--cleansing, attunement, and unification--addresses these challenges by (1) removing background frames, (2) leveraging large language models to extract clinical attributes for fine-grained contrastive learning, and (3) employing patient-level cross-attention to resolve multi-polyp ambiguities. Extensive experiments demonstrate that Endo-CLIP significantly outperforms state-of-the-art pre-training methods in zero-shot and few-shot polyp detection and classification, paving the way for more accurate and clinically relevant endoscopic analysis.