🤖 AI Summary
This work addresses the critical bottleneck in medical diagnosis: the absence of methods for dynamically generating high-fidelity images from clinical text. We propose the first dual-task text-to-image framework tailored for gastrointestinal imaging—comprising Image Synthesis (IS) and Optimal Prompt Generation (OPG). Methodologically, we systematically integrate fine-tuned Stable Diffusion, DreamBooth-based personalization, and LoRA-based low-rank adaptation, coupled with a CLIP text encoder to jointly optimize generation quality, class controllability, and diversity. Evaluated on multi-center data, our approach achieves FID = 0.064 and Inception Score = 2.327, significantly outperforming baseline models. Key contributions include: (1) transcending traditional static image analysis by enabling dynamic, natural-language-driven medical image synthesis; (2) establishing a scalable, high-precision prompt optimization mechanism; and (3) providing a reproducible technical pipeline and standardized evaluation benchmark for clinical-oriented generative AI.
📝 Abstract
The MEDVQA-GI challenge addresses the integration of AI-driven text-to-image generative models in medical diagnostics, aiming to enhance diagnostic capabilities through synthetic image generation. Existing methods primarily focus on static image analysis and lack the dynamic generation of medical imagery from textual descriptions. This study intends to partially close this gap by introducing a novel approach based on fine-tuned generative models to generate dynamic, scalable, and precise images from textual descriptions. Particularly, our system integrates fine-tuned Stable Diffusion and DreamBooth models, as well as Low-Rank Adaptation (LORA), to generate high-fidelity medical images. The problem is around two sub-tasks namely: image synthesis (IS) and optimal prompt production (OPG). The former creates medical images via verbal prompts, whereas the latter provides prompts that produce high-quality images in specified categories. The study emphasizes the limitations of traditional medical image generation methods, such as hand sketching, constrained datasets, static procedures, and generic models. Our evaluation measures showed that Stable Diffusion surpasses CLIP and DreamBooth + LORA in terms of producing high-quality, diversified images. Specifically, Stable Diffusion had the lowest Fr'echet Inception Distance (FID) scores (0.099 for single center, 0.064 for multi-center, and 0.067 for combined), indicating higher image quality. Furthermore, it had the highest average Inception Score (2.327 across all datasets), indicating exceptional diversity and quality. This advances the field of AI-powered medical diagnosis. Future research will concentrate on model refining, dataset augmentation, and ethical considerations for efficiently implementing these advances into clinical practice