🤖 AI Summary
Accurate benign-malignant classification of skin lesions from 3D full-body images captured under non-dermoscopic conditions (e.g., smartphone photography) remains challenging due to low contrast, occlusion, and severe class imbalance.
Method: We propose a segmentation-guided gradient-boosting hybrid framework: (1) U-Net guides lesion localization; (2) multi-scale visual features are extracted by fusing EVA02 ViT and our novel EdgeNeXtSAC architecture; (3) a diagnosis-driven relabeling strategy standardizes heterogeneous multi-source annotations into three clinically consistent classes, while Stable Diffusion synthesizes high-fidelity malignant lesion samples to mitigate class bias; (4) patient-level relational graph features are engineered and fed into XGBoost/GBDT for interpretable clinical decision-making.
Results: On the ISIC 2024 SLICE-3D benchmark, our method achieves a pAUC@TPR>80% of 0.1755—the highest score in this track—demonstrating substantial improvement in remote triage reliability under resource-constrained settings.
📝 Abstract
Skin cancer is among the most prevalent and life-threatening diseases worldwide, with early detection being critical to patient outcomes. This work presents a hybrid machine and deep learning-based approach for classifying malignant and benign skin lesions using the SLICE-3D dataset from ISIC 2024, which comprises 401,059 cropped lesion images extracted from 3D Total Body Photography (TBP), emulating non-dermoscopic, smartphone-like conditions. Our method combines vision transformers (EVA02) and our designed convolutional ViT hybrid (EdgeNeXtSAC) to extract robust features, employing a segmentation-assisted classification pipeline to enhance lesion localization. Predictions from these models are fused with a gradient-boosted decision tree (GBDT) ensemble enriched by engineered features and patient-specific relational metrics. To address class imbalance and improve generalization, we augment malignant cases with Stable Diffusion-generated synthetic lesions and apply a diagnosis-informed relabeling strategy to harmonize external datasets into a 3-class format. Using partial AUC (pAUC) above 80 percent true positive rate (TPR) as the evaluation metric, our approach achieves a pAUC of 0.1755 -- the highest among all configurations. These results underscore the potential of hybrid, interpretable AI systems for skin cancer triage in telemedicine and resource-constrained settings.