Uncertainty-Aware Explanations Through Probabilistic Self-Explainable Neural Networks

📅 2024-03-20
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
The black-box nature of deep neural networks hinders their trustworthy deployment in high-stakes applications. To address this, we propose Probabilistic Prototype-based Self-Explaining Neural Networks (Prob-PSENN), the first framework to elevate prototype learning from point estimation to a probabilistic modeling paradigm: prototypes are represented as distributions rather than fixed points, enabling joint optimization of prediction and explanation while explicitly quantifying explanation uncertainty. Our method integrates probabilistic deep learning, end-to-end differentiable prototype optimization, and Bayesian principles to identify uncertain predictions and provide calibrated, interpretable feedback. Evaluated on multiple benchmarks, Prob-PSENN generates more robust and semantically coherent explanations than deterministic counterparts, achieves predictive accuracy on par with non-probabilistic PSENN, and significantly enhances model trustworthiness and decision transparency in safety-critical domains.

Technology Category

Application Category

📝 Abstract
The lack of transparency of Deep Neural Networks continues to be a limitation that severely undermines their reliability and usage in high-stakes applications. Promising approaches to overcome such limitations are Prototype-Based Self-Explainable Neural Networks (PSENNs), whose predictions rely on the similarity between the input at hand and a set of prototypical representations of the output classes, offering therefore a deep, yet transparent-by-design, architecture. In this paper, we introduce a probabilistic reformulation of PSENNs, called Prob-PSENN, which replaces point estimates for the prototypes with probability distributions over their values. This provides not only a more flexible framework for an end-to-end learning of prototypes, but can also capture the explanatory uncertainty of the model, which is a missing feature in previous approaches. In addition, since the prototypes determine both the explanation and the prediction, Prob-PSENNs allow us to detect when the model is making uninformed or uncertain predictions, and to obtain valid explanations for them. Our experiments demonstrate that Prob-PSENNs provide more meaningful and robust explanations than their non-probabilistic counterparts, while remaining competitive in terms of predictive performance, thus enhancing the explainability and reliability of the models.
Problem

Research questions and friction points this paper is trying to address.

Improving transparency in Deep Neural Networks
Introducing uncertainty in model explanations
Enhancing reliability of high-stakes applications
Innovation

Methods, ideas, or system contributions that make the work stand out.

Probabilistic Self-Explainable Neural Networks
Uncertainty-aware model explanations
End-to-end prototype learning framework
🔎 Similar Papers
No similar papers found.
Jon Vadillo
Jon Vadillo
Intelligent Systems Group ISG, University of the Basque Country UPV/EHU
Adversarial Machine LearningDeep Learning
Roberto Santana
Roberto Santana
Intelligent Systems Group ISG, University of the Basque Country UPV/EHU
Estimation of Distribution AlgorithmsEvolutionary ComputationProbabilistic Graphical ModelsMachine Learning
J
J. A. Lozano
Department of Computer Science and Artificial Intelligence, University of the Basque Country UPV/EHU; Basque Center for Applied Mathematics (BCAM)
M
Marta Z. Kwiatkowska
Department of Computer Science, University of Oxford