🤖 AI Summary
In semi-supervised medical image segmentation, low-quality pseudo-labels and weak semantic consistency arise from the inability to jointly model heterogeneous voxel-level uncertainties from multiple sources. To address this, we propose a dual-evidence learning framework grounded in generalized evidence theory. Our key contributions are: (1) the novel Pister co-evidence fusion strategy, which unifies model-, data-, and structure-related uncertainties into a coherent evidential representation; (2) a quality-aware information volume metric (IVUM), enabling uncertainty-driven pseudo-label selection and weighting; and (3) an uncertainty-guided collaborative optimization mechanism coupled with an IVUM-weighted loss function. Evaluated on four benchmark medical imaging datasets, our method significantly improves segmentation accuracy, robustness, and generalization under both few-shot and noisy-label settings, achieving state-of-the-art performance.
📝 Abstract
Although existing semi-supervised image segmentation methods have achieved good performance, they cannot effectively utilize multiple sources of voxel-level uncertainty for targeted learning. Therefore, we propose two main improvements. First, we introduce a novel pignistic co-evidential fusion strategy using generalized evidential deep learning, extended by traditional D-S evidence theory, to obtain a more precise uncertainty measure for each voxel in medical samples. This assists the model in learning mixed labeled information and establishing semantic associations between labeled and unlabeled data. Second, we introduce the concept of information volume of mass function (IVUM) to evaluate the constructed evidence, implementing two evidential learning schemes. One optimizes evidential deep learning by combining the information volume of the mass function with original uncertainty measures. The other integrates the learning pattern based on the co-evidential fusion strategy, using IVUM to design a new optimization objective. Experiments on four datasets demonstrate the competitive performance of our method.