🤖 AI Summary
Insufficient interpretability of NLP models—particularly the lack of reliable, model-agnostic local explanations for discrete text inputs—hinders trust and transparency. To address this, we propose MASE, a model-agnostic explanation framework that estimates input feature importance by applying normalized linear Gaussian perturbations (NLGP) directly to the word embedding layer, thereby avoiding unstable discrete token perturbations. This design enhances explanation stability and robustness without requiring access to model internals or architectural assumptions. MASE is fully compatible with arbitrary black-box text classifiers. Extensive evaluation demonstrates that MASE significantly outperforms state-of-the-art model-agnostic methods—including LIME and SHAP—on metrics such as Delta Accuracy. By improving the traceability of prediction rationales and revealing model behavior more faithfully, MASE establishes a novel paradigm for trustworthy, post-hoc explanation in NLP.
📝 Abstract
Deep neural networks (DNNs) have made significant strides in Natural Language Processing (NLP), yet their interpretability remains elusive, particularly when evaluating their intricate decision-making processes. Traditional methods often rely on post-hoc interpretations, such as saliency maps or feature visualization, which might not be directly applicable to the discrete nature of word data in NLP. Addressing this, we introduce the Model-agnostic Saliency Estimation (MASE) framework. MASE offers local explanations for text-based predictive models without necessitating in-depth knowledge of a model's internal architecture. By leveraging Normalized Linear Gaussian Perturbations (NLGP) on the embedding layer instead of raw word inputs, MASE efficiently estimates input saliency. Our results indicate MASE's superiority over other model-agnostic interpretation methods, especially in terms of Delta Accuracy, positioning it as a promising tool for elucidating the operations of text-based models in NLP.