Improving the Faithfulness of Attention-based Explanations with
Task-specific Information for Text Classification
- URL: http://arxiv.org/abs/2105.02657v2
- Date: Fri, 7 May 2021 15:58:45 GMT
- Title: Improving the Faithfulness of Attention-based Explanations with
Task-specific Information for Text Classification
- Authors: George Chrysostomou and Nikolaos Aletras
- Abstract summary: We propose a new family of Task-Scaling (TaSc) mechanisms that learn task-specific non-contextualised information to scale the original attention weights.
TaSc consistently provides more faithful attention-based explanations compared to three widely-used interpretability techniques.
- Score: 9.147707153504117
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Neural network architectures in natural language processing often use
attention mechanisms to produce probability distributions over input token
representations. Attention has empirically been demonstrated to improve
performance in various tasks, while its weights have been extensively used as
explanations for model predictions. Recent studies (Jain and Wallace, 2019;
Serrano and Smith, 2019; Wiegreffe and Pinter, 2019) have showed that it cannot
generally be considered as a faithful explanation (Jacovi and Goldberg, 2020)
across encoders and tasks. In this paper, we seek to improve the faithfulness
of attention-based explanations for text classification. We achieve this by
proposing a new family of Task-Scaling (TaSc) mechanisms that learn
task-specific non-contextualised information to scale the original attention
weights. Evaluation tests for explanation faithfulness, show that the three
proposed variants of TaSc improve attention-based explanations across two
attention mechanisms, five encoders and five text classification datasets
without sacrificing predictive performance. Finally, we demonstrate that TaSc
consistently provides more faithful attention-based explanations compared to
three widely-used interpretability techniques.
Related papers
- Text Classification Based on Knowledge Graphs and Improved Attention
Mechanism [12.008192698720947]
The model operates at both character and word levels to deepen its understanding by integrating the concepts.
Its performance is demonstrated on datasets such as AGNews, Ohsumed, and TagMyNews, achieving accuracy of 75.1%, 58.7%, and 68.5% respectively.
arXiv Detail & Related papers (2024-01-07T22:20:55Z) - Generalized Robot 3D Vision-Language Model with Fast Rendering and Pre-Training Vision-Language Alignment [55.11291053011696]
This work presents a framework for dealing with 3D scene understanding when the labeled scenes are quite limited.
To extract knowledge for novel categories from the pre-trained vision-language models, we propose a hierarchical feature-aligned pre-training and knowledge distillation strategy.
In the limited reconstruction case, our proposed approach, termed WS3D++, ranks 1st on the large-scale ScanNet benchmark.
arXiv Detail & Related papers (2023-12-01T15:47:04Z) - Naturalness of Attention: Revisiting Attention in Code Language Models [3.756550107432323]
Language models for code such as CodeBERT offer the capability to learn advanced source code representation, but their opacity poses barriers to understanding of captured properties.
This study aims to shed some light on the previously ignored factors of the attention mechanism beyond the attention weights.
arXiv Detail & Related papers (2023-11-22T16:34:12Z) - Knowledge-Aware Prompt Tuning for Generalizable Vision-Language Models [64.24227572048075]
We propose a Knowledge-Aware Prompt Tuning (KAPT) framework for vision-language models.
Our approach takes inspiration from human intelligence in which external knowledge is usually incorporated into recognizing novel categories of objects.
arXiv Detail & Related papers (2023-08-22T04:24:45Z) - Perturbation-based Self-supervised Attention for Attention Bias in Text
Classification [31.144857032681905]
We propose a perturbation-based self-supervised attention approach to guide attention learning.
We add as much noise as possible to all the words in the sentence without changing their semantics and predictions.
Experimental results on three text classification tasks show that our approach can significantly improve the performance of current attention-based models.
arXiv Detail & Related papers (2023-05-25T03:18:18Z) - Fooling Explanations in Text Classifiers [42.49606659285249]
We introduceTextExplanationer (TEF), a novel explanation attack algorithm that alters text input samples imperceptibly.
TEF can significantly decrease the correlation between unchanged and perturbed input attributions.
We show that TEF perturbations are also effective in scenarios where the target model and explanation method are unknown.
arXiv Detail & Related papers (2022-06-07T10:58:08Z) - The Unreliability of Explanations in Few-Shot In-Context Learning [50.77996380021221]
We focus on two NLP tasks that involve reasoning over text, namely question answering and natural language inference.
We show that explanations judged as good by humans--those that are logically consistent with the input--usually indicate more accurate predictions.
We present a framework for calibrating model predictions based on the reliability of the explanations.
arXiv Detail & Related papers (2022-05-06T17:57:58Z) - Bayesian Attention Belief Networks [59.183311769616466]
Attention-based neural networks have achieved state-of-the-art results on a wide range of tasks.
This paper introduces Bayesian attention belief networks, which construct a decoder network by modeling unnormalized attention weights.
We show that our method outperforms deterministic attention and state-of-the-art attention in accuracy, uncertainty estimation, generalization across domains, and adversarial attacks.
arXiv Detail & Related papers (2021-06-09T17:46:22Z) - Variational Structured Attention Networks for Deep Visual Representation
Learning [49.80498066480928]
We propose a unified deep framework to jointly learn both spatial attention maps and channel attention in a principled manner.
Specifically, we integrate the estimation and the interaction of the attentions within a probabilistic representation learning framework.
We implement the inference rules within the neural network, thus allowing for end-to-end learning of the probabilistic and the CNN front-end parameters.
arXiv Detail & Related papers (2021-03-05T07:37:24Z) - Salience Estimation with Multi-Attention Learning for Abstractive Text
Summarization [86.45110800123216]
In the task of text summarization, salience estimation for words, phrases or sentences is a critical component.
We propose a Multi-Attention Learning framework which contains two new attention learning components for salience estimation.
arXiv Detail & Related papers (2020-04-07T02:38:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.