🤖 AI Summary
Existing unsupervised medical image registration methods rely primarily on intensity-based similarity metrics and lack explicit anatomical priors, limiting their accuracy and robustness. To address this, we propose the first unsupervised deformable registration framework that integrates anatomical knowledge from the Segment Anything Model (SAM). Specifically, we incorporate SAM-generated organ segmentation masks as explicit anatomical priors throughout the entire registration pipeline. We design a segmentation-prototype-based semantic alignment mechanism to enforce cross-image anatomical consistency, and introduce an edge-weighted contour-aware loss to improve robustness under ambiguous boundaries and complex deformations. Evaluated on multi-center datasets, our method achieves a 3.2% improvement in Dice score for liver and cardiac registration and reduces target registration error (TRE) by 28%, significantly outperforming state-of-the-art approaches. The source code is publicly available.
📝 Abstract
Medical image registration is a fundamental task in medical image analysis, aiming to establish spatial correspondences between paired images. However, existing unsupervised deformable registration methods rely solely on intensity-based similarity metrics, lacking explicit anatomical knowledge, which limits their accuracy and robustness. Vision foundation models, such as the Segment Anything Model (SAM), can generate high-quality segmentation masks that provide explicit anatomical structure knowledge, addressing the limitations of traditional methods that depend only on intensity similarity. Based on this, we propose a novel SAM-assisted registration framework incorporating prototype learning and contour awareness. The framework includes: (1) Explicit anatomical information injection, where SAM-generated segmentation masks are used as auxiliary inputs throughout training and testing to ensure the consistency of anatomical information; (2) Prototype learning, which leverages segmentation masks to extract prototype features and aligns prototypes to optimize semantic correspondences between images; and (3) Contour-aware loss, a contour-aware loss is designed that leverages the edges of segmentation masks to improve the model's performance in fine-grained deformation fields. Extensive experiments demonstrate that the proposed framework significantly outperforms existing methods across multiple datasets, particularly in challenging scenarios with complex anatomical structures and ambiguous boundaries. Our code is available at https://github.com/HaoXu0507/IPMI25-SAM-Assisted-Registration.