ICHOR: A Robust Representation Learning Approach for ASL CBF Maps with Self-Supervised Masked Autoencoders

📅 2026-03-05
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenge of model generalization in arterial spin labeling (ASL) cerebral blood flow (CBF) imaging, which is hindered by multicenter protocol heterogeneity, variable image quality, and scarce annotations. To overcome these limitations, we propose a self-supervised pretraining approach based on a three-dimensional masked autoencoder, uniquely integrating 3D masked image modeling with a Vision Transformer architecture. Leveraging the largest ASL CBF dataset to date—comprising 11,405 subjects—our method learns highly transferable representations. It significantly outperforms existing self-supervised neuroimaging approaches across three diagnostic classification tasks and one image quality control task, demonstrating markedly improved cross-center generalization. The pretrained weights and source code will be publicly released to facilitate further research.

Technology Category

Application Category

📝 Abstract
Arterial spin labeling (ASL) perfusion MRI allows direct quantification of regional cerebral blood flow (CBF) without exogenous contrast, enabling noninvasive measurements that can be repeated without constraints imposed by contrast injection. ASL is increasingly acquired in research studies and clinical MRI protocols. Building on successes in structural imaging, recent efforts have implemented deep learning based methods to improve image quality, enable automated quality control, and derive robust quantitative and predictive biomarkers with ASL derived CBF. However, progress has been limited by variable image quality, substantial inter-site, vendor and protocol differences, and limited availability of labeled datasets needed to train models that generalize across cohorts. To address these challenges, we introduce ICHOR, a self supervised pre-training approach for ASL CBF maps that learns transferable representations using 3D masked autoencoders. ICHOR is pretrained via masked image modeling using a Vision Transformer backbone and can be used as a general-purpose encoder for downstream ASL tasks. For pre-training, we curated one of the largest ASL datasets to date, comprising 11,405 ASL CBF scans from 14 studies spanning multiple sites and acquisition protocols. We evaluated the pre-trained ICHOR encoder on three downstream diagnostic classification tasks and one ASL CBF map quality prediction regression task. Across all evaluations, ICHOR outperformed existing neuroimaging self-supervised pre-training methods adapted to ASL. Pre-trained weights and code will be made publicly available.
Problem

Research questions and friction points this paper is trying to address.

Arterial Spin Labeling
Cerebral Blood Flow
Image Quality
Domain Generalization
Labeled Data Scarcity
Innovation

Methods, ideas, or system contributions that make the work stand out.

self-supervised learning
masked autoencoder
ASL CBF
Vision Transformer
multi-site generalization
🔎 Similar Papers
No similar papers found.
X
Xavier Beltran-Urbano
University of Pennsylvania, Philadelphia, USA
Y
Yiran Li
University of Maryland School of Medicine, Baltimore, USA
X
Xinglin Zeng
University of Maryland School of Medicine, Baltimore, USA
K
Katie R. Jobson
University of Pennsylvania, Philadelphia, USA
Manuel Taso
Manuel Taso
University of Pennsylvania
Magnetic Resonance Imaging
C
Christopher A. Brown
University of Pennsylvania, Philadelphia, USA
D
David A. Wolk
University of Pennsylvania, Philadelphia, USA
C
Corey T. McMillan
University of Pennsylvania, Philadelphia, USA
I
Ilya M. Nashrallah
University of Pennsylvania, Philadelphia, USA
P
Paul A. Yushkevich
University of Pennsylvania, Philadelphia, USA
Ze Wang
Ze Wang
University of Maryland School of Medicine
Arterial spin labelingbrain entropy mappingAlzheimer's Diseaseadolescence
J
John A. Detre
University of Pennsylvania, Philadelphia, USA
S
Sudipto Dolui
University of Pennsylvania, Philadelphia, USA