Learning to ignore: rethinking attention in CNNs
- URL: http://arxiv.org/abs/2111.05684v1
- Date: Wed, 10 Nov 2021 13:47:37 GMT
- Title: Learning to ignore: rethinking attention in CNNs
- Authors: Firas Laakom, Kateryna Chumachenko, Jenni Raitoharju, Alexandros
Iosifidis, and Moncef Gabbouj
- Abstract summary: We propose to reformulate the attention mechanism in CNNs to learn to ignore instead of learning to attend.
Specifically, we propose to explicitly learn irrelevant information in the scene and suppress it in the produced representation.
- Score: 87.01305532842878
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently, there has been an increasing interest in applying attention
mechanisms in Convolutional Neural Networks (CNNs) to solve computer vision
tasks. Most of these methods learn to explicitly identify and highlight
relevant parts of the scene and pass the attended image to further layers of
the network. In this paper, we argue that such an approach might not be
optimal. Arguably, explicitly learning which parts of the image are relevant is
typically harder than learning which parts of the image are less relevant and,
thus, should be ignored. In fact, in vision domain, there are many
easy-to-identify patterns of irrelevant features. For example, image regions
close to the borders are less likely to contain useful information for a
classification task. Based on this idea, we propose to reformulate the
attention mechanism in CNNs to learn to ignore instead of learning to attend.
Specifically, we propose to explicitly learn irrelevant information in the
scene and suppress it in the produced representation, keeping only important
attributes. This implicit attention scheme can be incorporated into any
existing attention mechanism. In this work, we validate this idea using two
recent attention methods Squeeze and Excitation (SE) block and Convolutional
Block Attention Module (CBAM). Experimental results on different datasets and
model architectures show that learning to ignore, i.e., implicit attention,
yields superior performance compared to the standard approaches.
Related papers
- SR-GNN: Spatial Relation-aware Graph Neural Network for Fine-Grained
Image Categorization [24.286426387100423]
We propose a method that captures subtle changes by aggregating context-aware features from most relevant image-regions.
Our approach is inspired by the recent advancement in self-attention and graph neural networks (GNNs)
It outperforms the state-of-the-art approaches by a significant margin in recognition accuracy.
arXiv Detail & Related papers (2022-09-05T19:43:15Z) - Visual Attention Network [90.0753726786985]
We propose a novel large kernel attention (LKA) module to enable self-adaptive and long-range correlations in self-attention.
We also introduce a novel neural network based on LKA, namely Visual Attention Network (VAN)
VAN outperforms the state-of-the-art vision transformers and convolutional neural networks with a large margin in extensive experiments.
arXiv Detail & Related papers (2022-02-20T06:35:18Z) - Where to Look: A Unified Attention Model for Visual Recognition with
Reinforcement Learning [5.247711598719703]
We propose to unify the top-down and bottom-up attention together for recurrent visual attention.
Our model exploits the image pyramids and Q-learning to select regions of interests in the top-down attention mechanism.
We train our model in an end-to-end reinforcement learning framework, and evaluate our method on visual classification tasks.
arXiv Detail & Related papers (2021-11-13T18:44:50Z) - Information Bottleneck Approach to Spatial Attention Learning [21.083618550304703]
The selective visual attention mechanism in the human visual system (HVS) restricts the amount of information to reach visual awareness for perceiving natural scenes.
This kind of selectivity acts as an 'Information Bottleneck (IB)', which seeks a trade-off between information compression and predictive accuracy.
We propose an IB-inspired spatial attention module for deep neural networks (DNNs) built for visual recognition.
arXiv Detail & Related papers (2021-08-07T10:35:32Z) - The Mind's Eye: Visualizing Class-Agnostic Features of CNNs [92.39082696657874]
We propose an approach to visually interpret CNN features given a set of images by creating corresponding images that depict the most informative features of a specific layer.
Our method uses a dual-objective activation and distance loss, without requiring a generator network nor modifications to the original model.
arXiv Detail & Related papers (2021-01-29T07:46:39Z) - One Point is All You Need: Directional Attention Point for Feature
Learning [51.44837108615402]
We present a novel attention-based mechanism for learning enhanced point features for tasks such as point cloud classification and segmentation.
We show that our attention mechanism can be easily incorporated into state-of-the-art point cloud classification and segmentation networks.
arXiv Detail & Related papers (2020-12-11T11:45:39Z) - Mining Cross-Image Semantics for Weakly Supervised Semantic Segmentation [128.03739769844736]
Two neural co-attentions are incorporated into the classifier to capture cross-image semantic similarities and differences.
In addition to boosting object pattern learning, the co-attention can leverage context from other related images to improve localization map inference.
Our algorithm sets new state-of-the-arts on all these settings, demonstrating well its efficacy and generalizability.
arXiv Detail & Related papers (2020-07-03T21:53:46Z) - Focus Longer to See Better:Recursively Refined Attention for
Fine-Grained Image Classification [148.4492675737644]
Deep Neural Network has shown great strides in the coarse-grained image classification task.
In this paper, we try to focus on these marginal differences to extract more representative features.
Our network repetitively focuses on parts of images to spot small discriminative parts among the classes.
arXiv Detail & Related papers (2020-05-22T03:14:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.