Self-Supervised Tuning for Few-Shot Segmentation
- URL: http://arxiv.org/abs/2004.05538v2
- Date: Mon, 14 Dec 2020 02:52:44 GMT
- Title: Self-Supervised Tuning for Few-Shot Segmentation
- Authors: Kai Zhu, Wei Zhai, Zheng-Jun Zha, Yang Cao
- Abstract summary: Few-shot segmentation aims at assigning a category label to each image pixel with few annotated samples.
Existing meta-learning method tends to fail in generating category-specifically discriminative descriptor when the visual features extracted from support images are marginalized in embedding space.
This paper presents an adaptive framework tuning, in which the distribution of latent features across different episodes is dynamically adjusted based on a self-segmentation scheme.
- Score: 82.32143982269892
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Few-shot segmentation aims at assigning a category label to each image pixel
with few annotated samples. It is a challenging task since the dense prediction
can only be achieved under the guidance of latent features defined by sparse
annotations. Existing meta-learning method tends to fail in generating
category-specifically discriminative descriptor when the visual features
extracted from support images are marginalized in embedding space. To address
this issue, this paper presents an adaptive tuning framework, in which the
distribution of latent features across different episodes is dynamically
adjusted based on a self-segmentation scheme, augmenting category-specific
descriptors for label prediction. Specifically, a novel self-supervised
inner-loop is firstly devised as the base learner to extract the underlying
semantic features from the support image. Then, gradient maps are calculated by
back-propagating self-supervised loss through the obtained features, and
leveraged as guidance for augmenting the corresponding elements in embedding
space. Finally, with the ability to continuously learn from different episodes,
an optimization-based meta-learner is adopted as outer loop of our proposed
framework to gradually refine the segmentation results. Extensive experiments
on benchmark PASCAL-$5^{i}$ and COCO-$20^{i}$ datasets demonstrate the
superiority of our proposed method over state-of-the-art.
Related papers
- Hierarchical Dense Correlation Distillation for Few-Shot
Segmentation-Extended Abstract [47.85056124410376]
Few-shot semantic segmentation (FSS) aims to form class-agnostic models segmenting unseen classes with only a handful of annotations.
We design Hierarchically Decoupled Matching Network (HDMNet) mining pixel-level support correlation based on the transformer architecture.
We propose a matching module to reduce train-set overfitting and introduce correlation distillation leveraging semantic correspondence from coarse resolution to boost fine-grained segmentation.
arXiv Detail & Related papers (2023-06-27T08:10:20Z) - Reflection Invariance Learning for Few-shot Semantic Segmentation [53.20466630330429]
Few-shot semantic segmentation (FSS) aims to segment objects of unseen classes in query images with only a few annotated support images.
This paper proposes a fresh few-shot segmentation framework to mine the reflection invariance in a multi-view matching manner.
Experiments on both PASCAL-$5textiti$ and COCO-$20textiti$ datasets demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2023-06-01T15:14:58Z) - Few-Shot Rotation-Invariant Aerial Image Semantic Segmentation [53.47160182465171]
Few-shot aerial image segmentation is a challenging task that involves precisely parsing objects in query aerial images.
The authors propose a novel few-shot rotation-invariant aerial semantic segmentation network (FRINet)
Experiments demonstrate that FRINet achieves state-of-the-art performance in few-shot aerial semantic segmentation benchmark.
arXiv Detail & Related papers (2023-05-29T09:28:34Z) - Hierarchical Dense Correlation Distillation for Few-Shot Segmentation [46.696051965252934]
Few-shot semantic segmentation (FSS) aims to form class-agnostic models segmenting unseen classes with only a handful of annotations.
We design Hierarchically Decoupled Matching Network (HDMNet) mining pixel-level support correlation based on the transformer architecture.
We propose a matching module to reduce train-set overfitting and introduce correlation distillation leveraging semantic correspondence from coarse resolution to boost fine-grained segmentation.
arXiv Detail & Related papers (2023-03-26T08:13:12Z) - Self-Supervised One-Shot Learning for Automatic Segmentation of StyleGAN
Images [0.0]
We propose a framework for the automatic one-shot segmentation of synthetic images generated by a StyleGAN.
Our framework learns to segment synthetic images using a self-supervised contrastive clustering algorithm.
We also show the results of using the proposed one-shot learner in implementing BagGAN, a framework for producing synthetic baggage X-ray scans for threat detection.
arXiv Detail & Related papers (2023-03-10T01:04:27Z) - From colouring-in to pointillism: revisiting semantic segmentation
supervision [48.637031591058175]
We propose a pointillist approach for semantic segmentation annotation, where only point-wise yes/no questions are answered.
We collected and released 22.6M point labels over 4,171 classes on the Open Images dataset.
arXiv Detail & Related papers (2022-10-25T16:42:03Z) - LEAD: Self-Supervised Landmark Estimation by Aligning Distributions of
Feature Similarity [49.84167231111667]
Existing works in self-supervised landmark detection are based on learning dense (pixel-level) feature representations from an image.
We introduce an approach to enhance the learning of dense equivariant representations in a self-supervised fashion.
We show that having such a prior in the feature extractor helps in landmark detection, even under drastically limited number of annotations.
arXiv Detail & Related papers (2022-04-06T17:48:18Z) - SCNet: Enhancing Few-Shot Semantic Segmentation by Self-Contrastive
Background Prototypes [56.387647750094466]
Few-shot semantic segmentation aims to segment novel-class objects in a query image with only a few annotated examples.
Most of advanced solutions exploit a metric learning framework that performs segmentation through matching each pixel to a learned foreground prototype.
This framework suffers from biased classification due to incomplete construction of sample pairs with the foreground prototype only.
arXiv Detail & Related papers (2021-04-19T11:21:47Z) - Self-Guided and Cross-Guided Learning for Few-Shot Segmentation [12.899804391102435]
We propose a self-guided learning approach for few-shot segmentation.
By making an initial prediction for the annotated support image, the covered and uncovered foreground regions are encoded to the primary and auxiliary support vectors.
By aggregating both primary and auxiliary support vectors, better segmentation performances are obtained on query images.
arXiv Detail & Related papers (2021-03-30T07:36:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.