🤖 AI Summary
Deep classification models for Alzheimer’s disease (AD) based on R2* mapping suffer from limited interpretability, hindering clinical trust and robustness. Method: This study systematically investigates how preprocessing and training configurations affect decision strategies. We propose a novel analytical framework integrating Layer-wise Relevance Propagation (LRP) with spectral clustering—first applying spectral clustering to 3D structural analysis of LRP heatmaps—and complementing it with t-SNE visualization. Contribution/Results: Relevance-guided training significantly enhances cluster separation between AD and normal controls (NC) in the heatmap representation space; t-SNE confirms strong alignment with clinical labels (12.3% accuracy improvement). The framework successfully identifies multiple distinct, interpretable disease-discriminative patterns, demonstrating that interpretable modeling critically enhances both the robustness and clinical credibility of medical AI systems.
📝 Abstract
Deep learning models have shown strong performance in classifying Alzheimer's disease (AD) from R2* maps, but their decision-making remains opaque, raising concerns about interpretability. Previous studies suggest biases in model decisions, necessitating further analysis. This study uses Layer-wise Relevance Propagation (LRP) and spectral clustering to explore classifier decision strategies across preprocessing and training configurations using R2* maps. We trained a 3D convolutional neural network on R2* maps, generating relevance heatmaps via LRP and applied spectral clustering to identify dominant patterns. t-Stochastic Neighbor Embedding (t-SNE) visualization was used to assess clustering structure. Spectral clustering revealed distinct decision patterns, with the relevance-guided model showing the clearest separation between AD and normal control (NC) cases. The t-SNE visualization confirmed that this model aligned heatmap groupings with the underlying subject groups. Our findings highlight the significant impact of preprocessing and training choices on deep learning models trained on R2* maps, even with similar performance metrics. Spectral clustering offers a structured method to identify classification strategy differences, emphasizing the importance of explainability in medical AI.