🤖 AI Summary
Tsetlin Machines (TMs) lack intrinsic uncertainty quantification and suffer from limited interpretability. Method: We propose a learning-dynamics-based internal probability scoring mechanism that leverages statistical properties of rule activation trajectories—without external calibration or post-hoc processing—to generate distribution-consistent probabilistic scores. Contribution/Results: This mechanism naturally characterizes prediction confidence and significantly improves out-of-distribution (OOD) detection. Unlike neural networks, which often overconfidently extrapolate, our approach identifies low-confidence predictions by analyzing the logical evolution of TM rules, enhancing transparency and trustworthiness. The probabilistic scores are inherently visualizable for interpretation. Empirical evaluation on synthetic data confirms score calibration, while experiments on CIFAR-10 expose structural limitations of current TM architectures. Overall, this work advances interpretable AI by establishing a principled, self-contained uncertainty-aware framework for logic-based learning.
📝 Abstract
Data modeling using Tsetlin machines (TMs) is all about building logical rules from the data features. The decisions of the model are based on a combination of these logical rules. Hence, the model is fully transparent and it is possible to get explanations of its predictions. In this paper, we present a probability score for TM predictions and develop new techniques for uncertainty quantification to increase the explainability further. The probability score is an inherent property of any TM variant and is derived through an analysis of the TM learning dynamics. Simulated data is used to show a clear connection between the learned TM probability scores and the underlying probabilities of the data. A visualization of the probability scores also reveals that the TM is less confident in its predictions outside the training data domain, which contrasts the typical extrapolation phenomenon found in Artificial Neural Networks. The paper concludes with an application of the uncertainty quantification techniques on an image classification task using the CIFAR-10 dataset, where they provide new insights and suggest possible improvements to current TM image classification models.