Prediction Calibration for Generalized Few-shot Semantic Segmentation
- URL: http://arxiv.org/abs/2210.08290v1
- Date: Sat, 15 Oct 2022 13:30:12 GMT
- Title: Prediction Calibration for Generalized Few-shot Semantic Segmentation
- Authors: Zhihe Lu, Sen He, Da Li, Yi-Zhe Song, Tao Xiang
- Abstract summary: Generalized Few-shot Semantic (GFSS) aims to segment each image pixel into either base classes with abundant training examples or novel classes with only a handful of (e.g., 1-5) training images per class.
We build a cross-attention module that guides the classifier's final prediction using the fused multi-level features.
Our PCN outperforms the state-the-art alternatives by large margins.
- Score: 101.69940565204816
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Generalized Few-shot Semantic Segmentation (GFSS) aims to segment each image
pixel into either base classes with abundant training examples or novel classes
with only a handful of (e.g., 1-5) training images per class. Compared to the
widely studied Few-shot Semantic Segmentation FSS, which is limited to
segmenting novel classes only, GFSS is much under-studied despite being more
practical. Existing approach to GFSS is based on classifier parameter fusion
whereby a newly trained novel class classifier and a pre-trained base class
classifier are combined to form a new classifier. As the training data is
dominated by base classes, this approach is inevitably biased towards the base
classes. In this work, we propose a novel Prediction Calibration Network PCN to
address this problem. Instead of fusing the classifier parameters, we fuse the
scores produced separately by the base and novel classifiers. To ensure that
the fused scores are not biased to either the base or novel classes, a new
Transformer-based calibration module is introduced. It is known that the
lower-level features are useful of detecting edge information in an input image
than higher-level features. Thus, we build a cross-attention module that guides
the classifier's final prediction using the fused multi-level features.
However, transformers are computationally demanding. Crucially, to make the
proposed cross-attention module training tractable at the pixel level, this
module is designed based on feature-score cross-covariance and episodically
trained to be generalizable at inference time. Extensive experiments on
PASCAL-$5^{i}$ and COCO-$20^{i}$ show that our PCN outperforms the
state-the-the-art alternatives by large margins.
Related papers
- Class Similarity Transition: Decoupling Class Similarities and Imbalance from Generalized Few-shot Segmentation [17.33292771556997]
This paper focuses on the relevance between base and novel classes, and improves Generalized Few-shot (GFSS)
We first propose a similarity transition matrix to guide the learning of novel classes with base class knowledge.
We then leverage the Label-Distribution-Aware Margin (LDAM) loss and Transductive Inference to the GFSS task to address the problem of class imbalance.
arXiv Detail & Related papers (2024-04-08T00:13:05Z) - Rethinking Few-shot 3D Point Cloud Semantic Segmentation [62.80639841429669]
This paper revisits few-shot 3D point cloud semantic segmentation (FS-PCS)
We focus on two significant issues in the state-of-the-art: foreground leakage and sparse point distribution.
To address these issues, we introduce a standardized FS-PCS setting, upon which a new benchmark is built.
arXiv Detail & Related papers (2024-03-01T15:14:47Z) - Neural Collapse Inspired Feature-Classifier Alignment for Few-Shot Class
Incremental Learning [120.53458753007851]
Few-shot class-incremental learning (FSCIL) has been a challenging problem as only a few training samples are accessible for each novel class in the new sessions.
We deal with this misalignment dilemma in FSCIL inspired by the recently discovered phenomenon named neural collapse.
We propose a neural collapse inspired framework for FSCIL. Experiments on the miniImageNet, CUB-200, and CIFAR-100 datasets demonstrate that our proposed framework outperforms the state-of-the-art performances.
arXiv Detail & Related papers (2023-02-06T18:39:40Z) - Class-Incremental Learning with Strong Pre-trained Models [97.84755144148535]
Class-incremental learning (CIL) has been widely studied under the setting of starting from a small number of classes (base classes)
We explore an understudied real-world setting of CIL that starts with a strong model pre-trained on a large number of base classes.
Our proposed method is robust and generalizes to all analyzed CIL settings.
arXiv Detail & Related papers (2022-04-07T17:58:07Z) - Learning What Not to Segment: A New Perspective on Few-Shot Segmentation [63.910211095033596]
Recently few-shot segmentation (FSS) has been extensively developed.
This paper proposes a fresh and straightforward insight to alleviate the problem.
In light of the unique nature of the proposed approach, we also extend it to a more realistic but challenging setting.
arXiv Detail & Related papers (2022-03-15T03:08:27Z) - Exploring Category-correlated Feature for Few-shot Image Classification [27.13708881431794]
We present a simple yet effective feature rectification method by exploring the category correlation between novel and base classes as the prior knowledge.
The proposed approach consistently obtains considerable performance gains on three widely used benchmarks.
arXiv Detail & Related papers (2021-12-14T08:25:24Z) - Variational Feature Disentangling for Fine-Grained Few-Shot
Classification [30.350307891161865]
Fine-grained few-shot recognition often suffers from the problem of training data scarcity for novel categories.
In this paper, we focus one enlarging the intra-class variance of the unseen class to improve few-shot classification performance.
arXiv Detail & Related papers (2020-10-07T08:13:42Z) - Pre-training Is (Almost) All You Need: An Application to Commonsense
Reasoning [61.32992639292889]
Fine-tuning of pre-trained transformer models has become the standard approach for solving common NLP tasks.
We introduce a new scoring method that casts a plausibility ranking task in a full-text format.
We show that our method provides a much more stable training phase across random restarts.
arXiv Detail & Related papers (2020-04-29T10:54:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.