🤖 AI Summary
This paper investigates classifiers’ ability to estimate their own prediction loss. Addressing the central question—“Can predictive distribution entropy reliably reflect actual loss?”—we propose and rigorously prove a computational equivalence between nontrivial loss prediction capability and multi-group calibration: a loss predictor outperforms the classifier’s self-estimate if and only if the latter violates multi-group calibration. This equivalence establishes, for the first time, a fundamental connection between loss prediction and calibration failure, implying that improving loss prediction is equivalent to detecting and correcting calibration biases. Our theoretical analysis integrates information theory, statistical learning theory, and the formal definition of multi-group calibration. Empirical modeling further confirms a robust positive correlation between loss prediction error and calibration violation magnitude. The work introduces a novel paradigm for uncertainty quantification and model diagnostics grounded in calibration-aware loss estimation.
📝 Abstract
Given a predictor and a loss function, how well can we predict the loss that the predictor will incur on an input? This is the problem of loss prediction, a key computational task associated with uncertainty estimation for a predictor. In a classification setting, a predictor will typically predict a distribution over labels and hence have its own estimate of the loss that it will incur, given by the entropy of the predicted distribution. Should we trust this estimate? In other words, when does the predictor know what it knows and what it does not know? In this work we study the theoretical foundations of loss prediction. Our main contribution is to establish tight connections between nontrivial loss prediction and certain forms of multicalibration, a multigroup fairness notion that asks for calibrated predictions across computationally identifiable subgroups. Formally, we show that a loss predictor that is able to improve on the self-estimate of a predictor yields a witness to a failure of multicalibration, and vice versa. This has the implication that nontrivial loss prediction is in effect no easier or harder than auditing for multicalibration. We support our theoretical results with experiments that show a robust positive correlation between the multicalibration error of a predictor and the efficacy of training a loss predictor.