🤖 AI Summary
Existing model-agnostic post-hoc attribution methods lack a systematic, quantitative theoretical framework. Method: This paper proposes TaylorPODA, a Taylor-expansion-based attribution theory framework. Contribution/Results: First, it formally defines three rigorous post-hoc attribution meta-principles—fidelity, federated compatibility, and zero-difference. Second, it introduces a task-adaptive mechanism for dynamic attribution calibration across diverse tasks. Third, it integrates meta-learning with visualization-based explanation techniques to balance theoretical rigor and practical flexibility. Experiments on multiple benchmark datasets demonstrate that TaylorPODA achieves or surpasses state-of-the-art performance in attribution fidelity and interpretability. Its visualizations are underpinned by stronger theoretical guarantees, significantly enhancing the credibility and deployment robustness of post-hoc explanations for opaque models.
📝 Abstract
Existing post-hoc model-agnostic methods generate external explanations for opaque models, primarily by locally attributing the model output to its input features. However, they often lack an explicit and systematic framework for quantifying the contribution of individual features. Building on the Taylor expansion framework introduced by Deng et al. (2024) to unify existing local attribution methods, we propose a rigorous set of postulates -- "precision", "federation", and "zero-discrepancy" -- to govern Taylor term-specific attribution. Guided by these postulates, we introduce TaylorPODA (Taylor expansion-derived imPortance-Order aDapted Attribution), which incorporates an additional "adaptation" property. This property enables alignment with task-specific goals, especially in post-hoc settings lacking ground-truth explanations. Empirical evaluations demonstrate that TaylorPODA achieves competitive results against baseline methods, providing principled and visualization-friendly explanations. This work represents a step toward the trustworthy deployment of opaque models by offering explanations with stronger theoretical grounding.