🤖 AI Summary
Image classification models suffer from poor interpretability, and pretrained representations are difficult to decompose into semantically meaningful components. Method: This paper proposes an information-bottleneck-based prototype disentanglement method that adaptively decomposes the final-layer features into interpretable, atomic-level prototype parts. It extends the prototype explanation paradigm—previously limited to small-scale settings—to large-scale, complex scenarios such as ImageNet, and unifies the flexibility of post-hoc explanation with the conceptual modeling capability of self-explaining networks, supporting both ViT and CNN backbones. Contribution/Results: By jointly optimizing feature disentanglement and enforcing information compression constraints, the method significantly improves explanation fidelity and concept consistency across multiple benchmark datasets. User studies confirm that the generated prototypes exhibit high human interpretability and practical utility in clinical/diagnostic contexts, providing fine-grained, verifiable semantic grounding for model decisions.
📝 Abstract
In this work, we introduce InfoDisent, a hybrid approach to explainability based on the information bottleneck principle. InfoDisent enables the disentanglement of information in the final layer of any pretrained model into atomic concepts, which can be interpreted as prototypical parts. This approach merges the flexibility of post-hoc methods with the concept-level modeling capabilities of self-explainable neural networks, such as ProtoPNets. We demonstrate the effectiveness of InfoDisent through computational experiments and user studies across various datasets using modern backbones such as ViTs and convolutional networks. Notably, InfoDisent generalizes the prototypical parts approach to novel domains (ImageNet).