Few-Shot Object Detection via Association and DIscrimination
- URL: http://arxiv.org/abs/2111.11656v1
- Date: Tue, 23 Nov 2021 05:04:06 GMT
- Title: Few-Shot Object Detection via Association and DIscrimination
- Authors: Yuhang Cao, Jiaqi Wang, Ying Jin, Tong Wu, Kai Chen, Ziwei Liu, Dahua
Lin
- Abstract summary: Few-shot object detection via Association and DIscrimination builds up a discriminative feature space for each novel class with two integral steps.
Experiments on Pascal VOC and MS-COCO datasets demonstrate FADI achieves new SOTA performance, significantly improving the baseline in any shot/split by +18.7.
- Score: 83.8472428718097
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Object detection has achieved substantial progress in the last decade.
However, detecting novel classes with only few samples remains challenging,
since deep learning under low data regime usually leads to a degraded feature
space. Existing works employ a holistic fine-tuning paradigm to tackle this
problem, where the model is first pre-trained on all base classes with abundant
samples, and then it is used to carve the novel class feature space.
Nonetheless, this paradigm is still imperfect. Durning fine-tuning, a novel
class may implicitly leverage the knowledge of multiple base classes to
construct its feature space, which induces a scattered feature space, hence
violating the inter-class separability. To overcome these obstacles, we propose
a two-step fine-tuning framework, Few-shot object detection via Association and
DIscrimination (FADI), which builds up a discriminative feature space for each
novel class with two integral steps. 1) In the association step, in contrast to
implicitly leveraging multiple base classes, we construct a compact novel class
feature space via explicitly imitating a specific base class feature space.
Specifically, we associate each novel class with a base class according to
their semantic similarity. After that, the feature space of a novel class can
readily imitate the well-trained feature space of the associated base class. 2)
In the discrimination step, to ensure the separability between the novel
classes and associated base classes, we disentangle the classification branches
for base and novel classes. To further enlarge the inter-class separability
between all classes, a set-specialized margin loss is imposed. Extensive
experiments on Pascal VOC and MS-COCO datasets demonstrate FADI achieves new
SOTA performance, significantly improving the baseline in any shot/split by
+18.7. Notably, the advantage is most announced on extremely few-shot
scenarios.
Related papers
- Covariance-based Space Regularization for Few-shot Class Incremental Learning [25.435192867105552]
Few-shot Class Incremental Learning (FSCIL) requires the model to continually learn new classes with limited labeled data.
Due to the limited data in incremental sessions, models are prone to overfitting new classes and suffering catastrophic forgetting of base classes.
Recent advancements resort to prototype-based approaches to constrain the base class distribution and learn discriminative representations of new classes.
arXiv Detail & Related papers (2024-11-02T08:03:04Z) - Semantic Enhanced Few-shot Object Detection [37.715912401900745]
We propose a fine-tuning based FSOD framework that utilizes semantic embeddings for better detection.
Our method allows each novel class to construct a compact feature space without being confused with similar base classes.
arXiv Detail & Related papers (2024-06-19T12:40:55Z) - Delve into Base-Novel Confusion: Redundancy Exploration for Few-Shot Class-Incremental Learning [8.729669566501103]
Few-shot class-incremental learning aims to acquire knowledge from novel classes with limited samples while retaining information about base classes.
Existing methods usually cause confusion between base and novel classes, classifying novel-class samples into base classes.
We propose Redundancy Decoupling and Integration (RDI) to alleviate this confusion.
arXiv Detail & Related papers (2024-05-08T09:38:16Z) - Expandable Subspace Ensemble for Pre-Trained Model-Based Class-Incremental Learning [65.57123249246358]
We propose ExpAndable Subspace Ensemble (EASE) for PTM-based CIL.
We train a distinct lightweight adapter module for each new task, aiming to create task-specific subspaces.
Our prototype complement strategy synthesizes old classes' new features without using any old class instance.
arXiv Detail & Related papers (2024-03-18T17:58:13Z) - Open-Vocabulary Object Detection with Meta Prompt Representation and Instance Contrastive Optimization [63.66349334291372]
We propose a framework with Meta prompt and Instance Contrastive learning (MIC) schemes.
Firstly, we simulate a novel-class-emerging scenario to help the prompt that learns class and background prompts generalize to novel classes.
Secondly, we design an instance-level contrastive strategy to promote intra-class compactness and inter-class separation, which benefits generalization of the detector to novel class objects.
arXiv Detail & Related papers (2024-03-14T14:25:10Z) - Few-Shot Class-Incremental Learning via Training-Free Prototype
Calibration [67.69532794049445]
We find a tendency for existing methods to misclassify the samples of new classes into base classes, which leads to the poor performance of new classes.
We propose a simple yet effective Training-frEE calibratioN (TEEN) strategy to enhance the discriminability of new classes.
arXiv Detail & Related papers (2023-12-08T18:24:08Z) - Learning with Fantasy: Semantic-Aware Virtual Contrastive Constraint for
Few-Shot Class-Incremental Learning [42.551377055029334]
Few-shot class-incremental learning (FSCIL) aims at learning to classify new classes continually from limited samples without forgetting the old classes.
We propose Semantic-Aware Virtual Contrastive model (SAVC), a novel method that facilitates separation between new classes and base classes.
Our SAVC significantly boosts base class separation and novel class generalization, achieving new state-of-the-art performance on the three widely-used FSCIL benchmark datasets.
arXiv Detail & Related papers (2023-04-02T01:51:24Z) - Activating the Discriminability of Novel Classes for Few-shot
Segmentation [48.542627940781095]
We propose to activate the discriminability of novel classes explicitly in both the feature encoding stage and the prediction stage for segmentation.
In the prediction stage for segmentation, we learn an Self-Refined Online Foreground-Background classifier (SROFB), which is able to refine itself using the high-confidence pixels of query image.
arXiv Detail & Related papers (2022-12-02T12:22:36Z) - Incremental Few-Shot Learning via Implanting and Compressing [13.122771115838523]
Incremental Few-Shot Learning requires a model to continually learn novel classes from only a few examples.
We propose a two-step learning strategy referred to as textbfImplanting and textbfCompressing.
Specifically, in the textbfImplanting step, we propose to mimic the data distribution of novel classes with the assistance of data-abundant base set.
In the textbf step, we adapt the feature extractor to precisely represent each novel class for enhancing intra-class compactness.
arXiv Detail & Related papers (2022-03-19T11:04:43Z) - Fine-grained Angular Contrastive Learning with Coarse Labels [72.80126601230447]
We introduce a novel 'Angular normalization' module that allows to effectively combine supervised and self-supervised contrastive pre-training.
This work will help to pave the way for future research on this new, challenging, and very practical topic of C2FS classification.
arXiv Detail & Related papers (2020-12-07T08:09:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.