Few-Shot Segmentation Without Meta-Learning: A Good Transductive
Inference Is All You Need?
- URL: http://arxiv.org/abs/2012.06166v2
- Date: Mon, 29 Mar 2021 21:03:58 GMT
- Title: Few-Shot Segmentation Without Meta-Learning: A Good Transductive
Inference Is All You Need?
- Authors: Malik Boudiaf, Hoel Kervadec, Ziko Imtiaz Masud, Pablo Piantanida,
Ismail Ben Ayed, Jose Dolz
- Abstract summary: We show that the way inference is performed in few-shot segmentation tasks has a substantial effect on performances.
We introduce a transductive inference for a given query image, leveraging the statistics of its unlabeled pixels.
We show that our method brings about 5% and 6% improvements over the state-of-the-art, in the 5- and 10-shot scenarios.
- Score: 34.95314059362982
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: We show that the way inference is performed in few-shot segmentation tasks
has a substantial effect on performances -- an aspect often overlooked in the
literature in favor of the meta-learning paradigm. We introduce a transductive
inference for a given query image, leveraging the statistics of its unlabeled
pixels, by optimizing a new loss containing three complementary terms: i) the
cross-entropy on the labeled support pixels; ii) the Shannon entropy of the
posteriors on the unlabeled query-image pixels; and iii) a global KL-divergence
regularizer based on the proportion of the predicted foreground. As our
inference uses a simple linear classifier of the extracted features, its
computational load is comparable to inductive inference and can be used on top
of any base training. Foregoing episodic training and using only standard
cross-entropy training on the base classes, our inference yields competitive
performances on standard benchmarks in the 1-shot scenarios. As the number of
available shots increases, the gap in performances widens: on PASCAL-5i, our
method brings about 5% and 6% improvements over the state-of-the-art, in the 5-
and 10-shot scenarios, respectively. Furthermore, we introduce a new setting
that includes domain shifts, where the base and novel classes are drawn from
different datasets. Our method achieves the best performances in this more
realistic setting. Our code is freely available online:
https://github.com/mboudiaf/RePRI-for-Few-Shot-Segmentation.
Related papers
- Intra-task Mutual Attention based Vision Transformer for Few-Shot Learning [12.5354658533836]
Humans possess remarkable ability to accurately classify new, unseen images after being exposed to only a few examples.
For artificial neural network models, determining the most relevant features for distinguishing between two images with limited samples presents a challenge.
We propose an intra-task mutual attention method for few-shot learning, that involves splitting the support and query samples into patches.
arXiv Detail & Related papers (2024-05-06T02:02:57Z) - Transductive Zero-Shot and Few-Shot CLIP [24.592841797020203]
This paper addresses the transductive zero-shot and few-shot CLIP classification challenge.
Inference is performed jointly across a mini-batch of unlabeled query samples, rather than treating each instance independently.
Our approach yields near 20% improvement in ImageNet accuracy over CLIP's zero-shot performance.
arXiv Detail & Related papers (2024-04-08T12:44:31Z) - Boosting Few-Shot Segmentation via Instance-Aware Data Augmentation and
Local Consensus Guided Cross Attention [7.939095881813804]
Few-shot segmentation aims to train a segmentation model that can fast adapt to a novel task for which only a few annotated images are provided.
We introduce an instance-aware data augmentation (IDA) strategy that augments the support images based on the relative sizes of the target objects.
The proposed IDA effectively increases the support set's diversity and promotes the distribution consistency between support and query images.
arXiv Detail & Related papers (2024-01-18T10:29:10Z) - Prediction Calibration for Generalized Few-shot Semantic Segmentation [101.69940565204816]
Generalized Few-shot Semantic (GFSS) aims to segment each image pixel into either base classes with abundant training examples or novel classes with only a handful of (e.g., 1-5) training images per class.
We build a cross-attention module that guides the classifier's final prediction using the fused multi-level features.
Our PCN outperforms the state-the-art alternatives by large margins.
arXiv Detail & Related papers (2022-10-15T13:30:12Z) - Transductive Decoupled Variational Inference for Few-Shot Classification [2.538209532048867]
Few-shot learning is an endeavour to transcend this capability down to machines.
We propose a novel variational inference network for few-shot classification (coined as TRIDENT)
We exploit information across both query and support images of a few-shot task using a novel built-in attention-based transductive feature extraction module (we call AttFEX)
arXiv Detail & Related papers (2022-08-22T19:27:09Z) - Learning What Not to Segment: A New Perspective on Few-Shot Segmentation [63.910211095033596]
Recently few-shot segmentation (FSS) has been extensively developed.
This paper proposes a fresh and straightforward insight to alleviate the problem.
In light of the unique nature of the proposed approach, we also extend it to a more realistic but challenging setting.
arXiv Detail & Related papers (2022-03-15T03:08:27Z) - Few-Shot Learning with Part Discovery and Augmentation from Unlabeled
Images [79.34600869202373]
We show that inductive bias can be learned from a flat collection of unlabeled images, and instantiated as transferable representations among seen and unseen classes.
Specifically, we propose a novel part-based self-supervised representation learning scheme to learn transferable representations.
Our method yields impressive results, outperforming the previous best unsupervised methods by 7.74% and 9.24%.
arXiv Detail & Related papers (2021-05-25T12:22:11Z) - SCNet: Enhancing Few-Shot Semantic Segmentation by Self-Contrastive
Background Prototypes [56.387647750094466]
Few-shot semantic segmentation aims to segment novel-class objects in a query image with only a few annotated examples.
Most of advanced solutions exploit a metric learning framework that performs segmentation through matching each pixel to a learned foreground prototype.
This framework suffers from biased classification due to incomplete construction of sample pairs with the foreground prototype only.
arXiv Detail & Related papers (2021-04-19T11:21:47Z) - A Few Guidelines for Incremental Few-Shot Segmentation [57.34237650765928]
Given a pretrained segmentation model and few images containing novel classes, our goal is to learn to segment novel classes while retaining the ability to segment previously seen ones.
We show how the main problems of end-to-end training in this scenario are.
i) the drift of the batch-normalization statistics toward novel classes that we can fix with batch renormalization and.
ii) the forgetting of old classes, that we can fix with regularization strategies.
arXiv Detail & Related papers (2020-11-30T20:45:56Z) - Pixel-Level Cycle Association: A New Perspective for Domain Adaptive
Semantic Segmentation [169.82760468633236]
We propose to build the pixel-level cycle association between source and target pixel pairs.
Our method can be trained end-to-end in one stage and introduces no additional parameters.
arXiv Detail & Related papers (2020-10-31T00:11:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.