A Probabilistic Segment Anything Model for Ambiguity-Aware Medical Image Segmentation

📅 2025-09-06
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing segmentation models (e.g., SAM) employ deterministic architectures, limiting their ability to capture intrinsic segmentation uncertainty arising from inter-expert annotation variability in medical imaging. To address this, we propose probabilistic SAM (pSAM), the first framework to integrate variational inference into the SAM architecture: it introduces a latent variable space that modulates prompt embeddings, jointly learning prior and posterior networks to explicitly model and efficiently sample from the segmentation output distribution. pSAM enables input-driven, diverse yet clinically plausible segmentations while providing calibrated uncertainty quantification. Evaluated on the LIDC-IDRI dataset, pSAM’s predicted segmentation distributions closely align with clinical expert disagreement—demonstrating significant improvements over state-of-the-art methods in both uncertainty calibration and segmentation diversity metrics. This work establishes a new paradigm for trustworthy, uncertainty-aware segmentation in medical imaging.

Technology Category

Application Category

📝 Abstract
Recent advances in promptable segmentation, such as the Segment Anything Model (SAM), have enabled flexible, high-quality mask generation across a wide range of visual domains. However, SAM and similar models remain fundamentally deterministic, producing a single segmentation per object per prompt, and fail to capture the inherent ambiguity present in many real-world tasks. This limitation is particularly troublesome in medical imaging, where multiple plausible segmentations may exist due to annotation uncertainty or inter-expert variability. In this paper, we introduce Probabilistic SAM, a probabilistic extension of SAM that models a distribution over segmentations conditioned on both the input image and prompt. By incorporating a latent variable space and training with a variational objective, our model learns to generate diverse and plausible segmentation masks reflecting the variability in human annotations. The architecture integrates a prior and posterior network into the SAM framework, allowing latent codes to modulate the prompt embeddings during inference. The latent space allows for efficient sampling during inference, enabling uncertainty-aware outputs with minimal overhead. We evaluate Probabilistic SAM on the public LIDC-IDRI lung nodule dataset and demonstrate its ability to produce diverse outputs that align with expert disagreement, outperforming existing probabilistic baselines on uncertainty-aware metrics. Our code is available at: https://github.com/tbwa233/Probabilistic-SAM/.
Problem

Research questions and friction points this paper is trying to address.

Addresses ambiguity in medical image segmentation
Models distribution over segmentations using latent variables
Generates diverse masks reflecting expert annotation variability
Innovation

Methods, ideas, or system contributions that make the work stand out.

Probabilistic extension of Segment Anything Model
Latent variable space with variational training
Generates diverse segmentation masks reflecting ambiguity
🔎 Similar Papers
No similar papers found.
Tyler Ward
Tyler Ward
Student Teaching Assistant, University of Kentucky
Computer visionMachine learningMedical imagingQuality engineering
A
Abdullah Imran
Department of Computer Science, University of Kentucky, Lexington, KY 40506, USA