🤖 AI Summary
Deep neural networks’ visual decision-making lacks interpretability, risking unwarranted trust and latent biases. To address this, we propose *Clustered Paths*: a method that clusters hierarchical neural activations to generate human-readable, layer-wise evolving decision paths—explicitly revealing the color, texture, and contextual visual concepts leveraged by the model at each depth. We introduce the first four-dimensional quantitative evaluation framework, assessing cognitive load, class alignment, prediction fidelity, and perturbation stability. Furthermore, we integrate large language models to generate high-level semantic concept paths, enabling out-of-distribution (OOD) anomaly detection. On CIFAR-10 and CelebA, our approach accurately identifies spurious correlations with >90% fidelity and 96% perturbation stability. It scales effectively to ImageNet-scale Vision Transformers (ViTs) and achieves efficient, reliable OOD detection.
📝 Abstract
While modern deep neural networks achieve impressive performance in vision tasks, they remain opaque in their decision processes, risking unwarranted trust, undetected biases and unexpected failures. We propose cluster paths, a post-hoc interpretability method that clusters activations at selected layers and represents each input as its sequence of cluster IDs. To assess these cluster paths, we introduce four metrics: path complexity (cognitive load), weighted-path purity (class alignment), decision-alignment faithfulness (predictive fidelity), and path agreement (stability under perturbations). In a spurious-cue CIFAR-10 experiment, cluster paths identify color-based shortcuts and collapse when the cue is removed. On a five-class CelebA hair-color task, they achieve 90% faithfulness and maintain 96% agreement under Gaussian noise without sacrificing accuracy. Scaling to a Vision Transformer pretrained on ImageNet, we extend cluster paths to concept paths derived from prompting a large language model on minimal path divergences. Finally, we show that cluster paths can serve as an effective out-of-distribution (OOD) detector, reliably flagging anomalous samples before the model generates over-confident predictions. Cluster paths uncover visual concepts, such as color palettes, textures, or object contexts, at multiple network depths, demonstrating that cluster paths scale to large vision models while generating concise and human-readable explanations.