Attention Meets Perturbations: Robust and Interpretable Attention with
Adversarial Training
- URL: http://arxiv.org/abs/2009.12064v2
- Date: Thu, 1 Jul 2021 02:31:22 GMT
- Title: Attention Meets Perturbations: Robust and Interpretable Attention with
Adversarial Training
- Authors: Shunsuke Kitada and Hitoshi Iyatomi
- Abstract summary: We propose a general training technique for natural language processing tasks, including AT for attention (Attention AT) and more interpretable AT for attention (Attention iAT)
The proposed techniques improved the prediction performance and the model interpretability by exploiting the mechanisms with AT.
- Score: 7.106986689736828
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Although attention mechanisms have been applied to a variety of deep learning
models and have been shown to improve the prediction performance, it has been
reported to be vulnerable to perturbations to the mechanism. To overcome the
vulnerability to perturbations in the mechanism, we are inspired by adversarial
training (AT), which is a powerful regularization technique for enhancing the
robustness of the models. In this paper, we propose a general training
technique for natural language processing tasks, including AT for attention
(Attention AT) and more interpretable AT for attention (Attention iAT). The
proposed techniques improved the prediction performance and the model
interpretability by exploiting the mechanisms with AT. In particular, Attention
iAT boosts those advantages by introducing adversarial perturbation, which
enhances the difference in the attention of the sentences. Evaluation
experiments with ten open datasets revealed that AT for attention mechanisms,
especially Attention iAT, demonstrated (1) the best performance in nine out of
ten tasks and (2) more interpretable attention (i.e., the resulting attention
correlated more strongly with gradient-based word importance) for all tasks.
Additionally, the proposed techniques are (3) much less dependent on
perturbation size in AT. Our code is available at
https://github.com/shunk031/attention-meets-perturbation
Related papers
- Towards Robust Semantic Segmentation against Patch-based Attack via Attention Refinement [68.31147013783387]
We observe that the attention mechanism is vulnerable to patch-based adversarial attacks.
In this paper, we propose a Robust Attention Mechanism (RAM) to improve the robustness of the semantic segmentation model.
arXiv Detail & Related papers (2024-01-03T13:58:35Z) - Perturbation-based Self-supervised Attention for Attention Bias in Text
Classification [31.144857032681905]
We propose a perturbation-based self-supervised attention approach to guide attention learning.
We add as much noise as possible to all the words in the sentence without changing their semantics and predictions.
Experimental results on three text classification tasks show that our approach can significantly improve the performance of current attention-based models.
arXiv Detail & Related papers (2023-05-25T03:18:18Z) - Attention mechanisms for physiological signal deep learning: which
attention should we take? [0.0]
We experimentally analyze four attention mechanisms (e.g., squeeze-and-excitation, non-local, convolutional block attention module, and multi-head self-attention) and three convolutional neural network (CNN) architectures.
We evaluate multiple combinations for performance and convergence of physiological signal deep learning model.
arXiv Detail & Related papers (2022-07-04T07:24:08Z) - Self-Supervised Implicit Attention: Guided Attention by The Model Itself [1.3406858660972554]
We propose Self-Supervised Implicit Attention (SSIA), a new approach that adaptively guides deep neural network models to gain attention by exploiting the properties of the models themselves.
SSIAA is a novel attention mechanism that does not require any extra parameters, computation, or memory access costs during inference.
Our implementation will be available on GitHub.
arXiv Detail & Related papers (2022-06-15T10:13:34Z) - Guiding Visual Question Answering with Attention Priors [76.21671164766073]
We propose to guide the attention mechanism using explicit linguistic-visual grounding.
This grounding is derived by connecting structured linguistic concepts in the query to their referents among the visual objects.
The resultant algorithm is capable of probing attention-based reasoning models, injecting relevant associative knowledge, and regulating the core reasoning process.
arXiv Detail & Related papers (2022-05-25T09:53:47Z) - Alignment Attention by Matching Key and Query Distributions [48.93793773929006]
This paper introduces alignment attention that explicitly encourages self-attention to match the distributions of the key and query within each head.
It is simple to convert any models with self-attention, including pre-trained ones, to the proposed alignment attention.
On a variety of language understanding tasks, we show the effectiveness of our method in accuracy, uncertainty estimation, generalization across domains, and robustness to adversarial attacks.
arXiv Detail & Related papers (2021-10-25T00:54:57Z) - Counterfactual Attention Learning for Fine-Grained Visual Categorization
and Re-identification [101.49122450005869]
We present a counterfactual attention learning method to learn more effective attention based on causal inference.
Specifically, we analyze the effect of the learned visual attention on network prediction.
We evaluate our method on a wide range of fine-grained recognition tasks.
arXiv Detail & Related papers (2021-08-19T14:53:40Z) - More Than Just Attention: Learning Cross-Modal Attentions with
Contrastive Constraints [63.08768589044052]
We propose Contrastive Content Re-sourcing ( CCR) and Contrastive Content Swapping ( CCS) constraints to address such limitation.
CCR and CCS constraints supervise the training of attention models in a contrastive learning manner without requiring explicit attention annotations.
Experiments on both Flickr30k and MS-COCO datasets demonstrate that integrating these attention constraints into two state-of-the-art attention-based models improves the model performance.
arXiv Detail & Related papers (2021-05-20T08:48:10Z) - Making Attention Mechanisms More Robust and Interpretable with Virtual
Adversarial Training for Semi-Supervised Text Classification [9.13755431537592]
We propose a new general training technique for attention mechanisms based on virtual adversarial training (VAT)
VAT can compute adversarial perturbations from unlabeled data in a semi-supervised setting for the attention mechanisms that have been reported in previous studies to be vulnerable to perturbations.
arXiv Detail & Related papers (2021-04-18T07:51:45Z) - Cost-effective Interactive Attention Learning with Neural Attention
Processes [79.8115563067513]
We propose a novel interactive learning framework which we refer to as Interactive Attention Learning (IAL)
IAL is prone to overfitting due to scarcity of human annotations, and requires costly retraining.
We tackle these challenges by proposing a sample-efficient attention mechanism and a cost-effective reranking algorithm for instances and features.
arXiv Detail & Related papers (2020-06-09T17:36:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.