Task-Oriented Channel Attention for Fine-Grained Few-Shot Classification
- URL: http://arxiv.org/abs/2308.00093v1
- Date: Fri, 28 Jul 2023 08:40:23 GMT
- Title: Task-Oriented Channel Attention for Fine-Grained Few-Shot Classification
- Authors: SuBeen Lee, WonJun Moon, Hyun Seok Seong, and Jae-Pil Heo
- Abstract summary: Task Discrepancy Maximization (TDM) is a task-oriented channel attention method tailored for fine-grained few-shot classification.
SAM highlights channels encoding class-wise discriminative features, while QAM assigns higher weights to object-relevant channels of the query.
Based on these submodules, TDM produces task-adaptive features by focusing on channels encoding class-discriminative details and possessed by the query.
- Score: 5.4352987210173955
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The difficulty of the fine-grained image classification mainly comes from a
shared overall appearance across classes. Thus, recognizing discriminative
details, such as eyes and beaks for birds, is a key in the task. However, this
is particularly challenging when training data is limited. To address this, we
propose Task Discrepancy Maximization (TDM), a task-oriented channel attention
method tailored for fine-grained few-shot classification with two novel modules
Support Attention Module (SAM) and Query Attention Module (QAM). SAM highlights
channels encoding class-wise discriminative features, while QAM assigns higher
weights to object-relevant channels of the query. Based on these submodules,
TDM produces task-adaptive features by focusing on channels encoding
class-discriminative details and possessed by the query at the same time, for
accurate class-sensitive similarity measure between support and query
instances. While TDM influences high-level feature maps by task-adaptive
calibration of channel-wise importance, we further introduce Instance Attention
Module (IAM) operating in intermediate layers of feature extractors to
instance-wisely highlight object-relevant channels, by extending QAM. The
merits of TDM and IAM and their complementary benefits are experimentally
validated in fine-grained few-shot classification tasks. Moreover, IAM is also
shown to be effective in coarse-grained and cross-domain few-shot
classifications.
Related papers
- Multiple Instance Verification [11.027466339522777]
We show that naive adaptations of attention-based multiple instance learning methods and standard verification methods are unsuitable for this setting.
Under the CAP framework, we propose two novel attention functions to address the challenge of distinguishing between highly similar instances in a target bag.
arXiv Detail & Related papers (2024-07-09T04:51:22Z) - Auxiliary Tasks Enhanced Dual-affinity Learning for Weakly Supervised
Semantic Segmentation [79.05949524349005]
We propose AuxSegNet+, a weakly supervised auxiliary learning framework to explore the rich information from saliency maps.
We also propose a cross-task affinity learning mechanism to learn pixel-level affinities from the saliency and segmentation feature maps.
arXiv Detail & Related papers (2024-03-02T10:03:21Z) - Enhancing Few-shot CLIP with Semantic-Aware Fine-Tuning [61.902254546858465]
Methods based on Contrastive Language-Image Pre-training have exhibited promising performance in few-shot adaptation tasks.
We propose fine-tuning the parameters of the attention pooling layer during the training process to encourage the model to focus on task-specific semantics.
arXiv Detail & Related papers (2023-11-08T05:18:57Z) - Task Discrepancy Maximization for Fine-grained Few-Shot Classification [6.0158981171030685]
Task Discrepancy Maximization (TDM) is a simple module for fine-grained few-shot classification.
Our objective is to localize the class-wise discriminative regions by highlighting channels encoding distinct information of the class.
arXiv Detail & Related papers (2022-07-04T12:54:58Z) - Continual Object Detection via Prototypical Task Correlation Guided
Gating Mechanism [120.1998866178014]
We present a flexible framework for continual object detection via pRotOtypical taSk corrElaTion guided gaTingAnism (ROSETTA)
Concretely, a unified framework is shared by all tasks while task-aware gates are introduced to automatically select sub-models for specific tasks.
Experiments on COCO-VOC, KITTI-Kitchen, class-incremental detection on VOC and sequential learning of four tasks show that ROSETTA yields state-of-the-art performance.
arXiv Detail & Related papers (2022-05-06T07:31:28Z) - Activation Modulation and Recalibration Scheme for Weakly Supervised
Semantic Segmentation [24.08326440298189]
We propose a novel activation modulation and recalibration scheme for weakly supervised semantic segmentation.
We show that AMR establishes a new state-of-the-art performance on the PASCAL VOC 2012 dataset.
Experiments also reveal that our scheme is plug-and-play and can be incorporated with other approaches to boost their performance.
arXiv Detail & Related papers (2021-12-16T16:26:14Z) - Leveraging Auxiliary Tasks with Affinity Learning for Weakly Supervised
Semantic Segmentation [88.49669148290306]
We propose a novel weakly supervised multi-task framework called AuxSegNet to leverage saliency detection and multi-label image classification as auxiliary tasks.
Inspired by their similar structured semantics, we also propose to learn a cross-task global pixel-level affinity map from the saliency and segmentation representations.
The learned cross-task affinity can be used to refine saliency predictions and propagate CAM maps to provide improved pseudo labels for both tasks.
arXiv Detail & Related papers (2021-07-25T11:39:58Z) - Semantic Attention and Scale Complementary Network for Instance
Segmentation in Remote Sensing Images [54.08240004593062]
We propose an end-to-end multi-category instance segmentation model, which consists of a Semantic Attention (SEA) module and a Scale Complementary Mask Branch (SCMB)
SEA module contains a simple fully convolutional semantic segmentation branch with extra supervision to strengthen the activation of interest instances on the feature map.
SCMB extends the original single mask branch to trident mask branches and introduces complementary mask supervision at different scales.
arXiv Detail & Related papers (2021-07-25T08:53:59Z) - Channel DropBlock: An Improved Regularization Method for Fine-Grained
Visual Classification [58.07257910065007]
Existing approaches mainly tackle this problem by introducing attention mechanisms to locate the discriminative parts or feature encoding approaches to extract the highly parameterized features in a weakly-supervised fashion.
In this work, we propose a lightweight yet effective regularization method named Channel DropBlock (CDB) in combination with two alternative correlation metrics, to address this problem.
arXiv Detail & Related papers (2021-06-07T09:03:02Z) - Fine-Grained Visual Classification via Simultaneously Learning of
Multi-regional Multi-grained Features [15.71408474557042]
Fine-grained visual classification is a challenging task that recognizes the sub-classes belonging to the same meta-class.
In this paper, we argue that mining multi-regional multi-grained features is precisely the key to this task.
Experimental results over four widely used fine-grained image classification datasets demonstrate the effectiveness of the proposed method.
arXiv Detail & Related papers (2021-01-31T03:46:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.