Task Discrepancy Maximization for Fine-grained Few-Shot Classification
- URL: http://arxiv.org/abs/2207.01376v1
- Date: Mon, 4 Jul 2022 12:54:58 GMT
- Title: Task Discrepancy Maximization for Fine-grained Few-Shot Classification
- Authors: SuBeen Lee, WonJun Moon, Jae-Pil Heo
- Abstract summary: Task Discrepancy Maximization (TDM) is a simple module for fine-grained few-shot classification.
Our objective is to localize the class-wise discriminative regions by highlighting channels encoding distinct information of the class.
- Score: 6.0158981171030685
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recognizing discriminative details such as eyes and beaks is important for
distinguishing fine-grained classes since they have similar overall
appearances. In this regard, we introduce Task Discrepancy Maximization (TDM),
a simple module for fine-grained few-shot classification. Our objective is to
localize the class-wise discriminative regions by highlighting channels
encoding distinct information of the class. Specifically, TDM learns
task-specific channel weights based on two novel components: Support Attention
Module (SAM) and Query Attention Module (QAM). SAM produces a support weight to
represent channel-wise discriminative power for each class. Still, since the
SAM is basically only based on the labeled support sets, it can be vulnerable
to bias toward such support set. Therefore, we propose QAM which complements
SAM by yielding a query weight that grants more weight to object-relevant
channels for a given query image. By combining these two weights, a class-wise
task-specific channel weight is defined. The weights are then applied to
produce task-adaptive feature maps more focusing on the discriminative details.
Our experiments validate the effectiveness of TDM and its complementary
benefits with prior methods in fine-grained few-shot classification.
Related papers
- There is no SAMantics! Exploring SAM as a Backbone for Visual Understanding Tasks [15.061599989448867]
The Segment Anything Model (SAM) was originally designed for label-agnostic mask generation.
We quantify SAM's semantic capabilities by comparing base image encoder efficacy under classification tasks.
Our findings reveal a significant lack of semantic discriminability in SAM feature representations.
arXiv Detail & Related papers (2024-11-22T17:00:18Z) - AlignSAM: Aligning Segment Anything Model to Open Context via Reinforcement Learning [61.666973416903005]
Segment Anything Model (SAM) has demonstrated its impressive generalization capabilities in open-world scenarios with the guidance of prompts.
We propose a novel framework, termed AlignSAM, designed for automatic prompting for aligning SAM to an open context.
arXiv Detail & Related papers (2024-06-01T16:21:39Z) - GoodSAM: Bridging Domain and Capacity Gaps via Segment Anything Model for Distortion-aware Panoramic Semantic Segmentation [22.344399402787644]
This paper tackles a novel yet challenging problem: how to transfer knowledge from the emerging Segment Anything Model (SAM)
We propose a framework, called GoodSAM, that introduces a teacher assistant (TA) to provide semantic information, integrated with SAM to generate ensemble logits.
Experiments on two benchmarks show that our GoodSAM achieves a remarkable +3.75% mIoU improvement over the state-of-the-art (SOTA) domain adaptation methods.
arXiv Detail & Related papers (2024-03-25T02:30:32Z) - PosSAM: Panoptic Open-vocabulary Segment Anything [58.72494640363136]
PosSAM is an open-vocabulary panoptic segmentation model that unifies the strengths of the Segment Anything Model (SAM) with the vision-native CLIP model in an end-to-end framework.
We introduce a Mask-Aware Selective Ensembling (MASE) algorithm that adaptively enhances the quality of generated masks and boosts the performance of open-vocabulary classification during inference for each image.
arXiv Detail & Related papers (2024-03-14T17:55:03Z) - SFC: Shared Feature Calibration in Weakly Supervised Semantic
Segmentation [28.846513129022803]
Image-level weakly supervised semantic segmentation has received increasing attention due to its low annotation cost.
Existing methods mainly rely on Class Mapping (CAM) to obtain pseudo-labels for training semantic segmentation models.
In this work, we are the first to demonstrate that long-tailed distribution in training data can cause the CAM calculated through weights over-activated for head classes and under-activated for tail classes due to the shared features among head- and tail- classes.
arXiv Detail & Related papers (2024-01-22T06:43:13Z) - Stable Segment Anything Model [79.9005670886038]
The Segment Anything Model (SAM) achieves remarkable promptable segmentation given high-quality prompts.
This paper presents the first comprehensive analysis on SAM's segmentation stability across a diverse spectrum of prompt qualities.
Our solution, termed Stable-SAM, offers several advantages: 1) improved SAM's segmentation stability across a wide range of prompt qualities, while 2) retaining SAM's powerful promptable segmentation efficiency and generality.
arXiv Detail & Related papers (2023-11-27T12:51:42Z) - Task-Oriented Channel Attention for Fine-Grained Few-Shot Classification [5.4352987210173955]
Task Discrepancy Maximization (TDM) is a task-oriented channel attention method tailored for fine-grained few-shot classification.
SAM highlights channels encoding class-wise discriminative features, while QAM assigns higher weights to object-relevant channels of the query.
Based on these submodules, TDM produces task-adaptive features by focusing on channels encoding class-discriminative details and possessed by the query.
arXiv Detail & Related papers (2023-07-28T08:40:23Z) - PatchMix Augmentation to Identify Causal Features in Few-shot Learning [55.64873998196191]
Few-shot learning aims to transfer knowledge learned from base with sufficient categories labelled data to novel categories with scarce known information.
We propose a novel data augmentation strategy dubbed as PatchMix that can break this spurious dependency.
We show that such an augmentation mechanism, different from existing ones, is able to identify the causal features.
arXiv Detail & Related papers (2022-11-29T08:41:29Z) - Boosting Few-shot Fine-grained Recognition with Background Suppression
and Foreground Alignment [53.401889855278704]
Few-shot fine-grained recognition (FS-FGR) aims to recognize novel fine-grained categories with the help of limited available samples.
We propose a two-stage background suppression and foreground alignment framework, which is composed of a background activation suppression (BAS) module, a foreground object alignment (FOA) module, and a local to local (L2L) similarity metric.
Experiments conducted on multiple popular fine-grained benchmarks demonstrate that our method outperforms the existing state-of-the-art by a large margin.
arXiv Detail & Related papers (2022-10-04T07:54:40Z) - Revisiting Unsupervised Meta-Learning: Amplifying or Compensating for
the Characteristics of Few-Shot Tasks [30.893785366366078]
We develop a practical approach towards few-shot image classification, where a visual recognition system is constructed with limited data.
We find that the base class set labels are not necessary, and discriminative embeddings could be meta-learned in an unsupervised manner.
Experiments on few-shot learning benchmarks verify our approaches outperform previous methods by a 4-10% performance gap.
arXiv Detail & Related papers (2020-11-30T10:08:35Z) - A Few-Shot Sequential Approach for Object Counting [63.82757025821265]
We introduce a class attention mechanism that sequentially attends to objects in the image and extracts their relevant features.
The proposed technique is trained on point-level annotations and uses a novel loss function that disentangles class-dependent and class-agnostic aspects of the model.
We present our results on a variety of object-counting/detection datasets, including FSOD and MS COCO.
arXiv Detail & Related papers (2020-07-03T18:23:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.