Task-Adaptive Feature Transformer for Few-Shot Segmentation
- URL: http://arxiv.org/abs/2010.11437v1
- Date: Thu, 22 Oct 2020 04:35:37 GMT
- Title: Task-Adaptive Feature Transformer for Few-Shot Segmentation
- Authors: Jun Seo, Young-Hyun Park, Sung-Whan Yoon, Jaekyun Moon
- Abstract summary: We propose a learnable module for few-shot segmentation, the task-adaptive feature transformer (TAFT)
TAFT linearly transforms task-specific high-level features to a set of task-agnostic features well-suited to the segmentation job.
Experiments on the PASCAL-$5i$ dataset confirm that this combination successfully adds few-shot learning capability to the segmentation algorithm.
- Score: 21.276981570672064
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Few-shot learning allows machines to classify novel classes using only a few
labeled samples. Recently, few-shot segmentation aiming at semantic
segmentation on low sample data has also seen great interest. In this paper, we
propose a learnable module for few-shot segmentation, the task-adaptive feature
transformer (TAFT). TAFT linearly transforms task-specific high-level features
to a set of task-agnostic features well-suited to the segmentation job. Using
this task-conditioned feature transformation, the model is shown to effectively
utilize the semantic information in novel classes to generate tight
segmentation masks. The proposed TAFT module can be easily plugged into
existing semantic segmentation algorithms to achieve few-shot segmentation
capability with only a few added parameters. We combine TAFT with Deeplab V3+,
a well-known segmentation architecture; experiments on the PASCAL-$5^i$ dataset
confirm that this combination successfully adds few-shot learning capability to
the segmentation algorithm, achieving the state-of-the-art few-shot
segmentation performance in some key representative cases.
Related papers
- OMG-Seg: Is One Model Good Enough For All Segmentation? [83.17068644513144]
OMG-Seg is a transformer-based encoder-decoder architecture with task-specific queries and outputs.
We show that OMG-Seg can support over ten distinct segmentation tasks and yet significantly reduce computational and parameter overhead.
arXiv Detail & Related papers (2024-01-18T18:59:34Z) - FreeSeg: Unified, Universal and Open-Vocabulary Image Segmentation [42.89720785573885]
FreeSeg is a generic framework to accomplish Unified, Universal and Open-Vocabulary Image.
We show that FreeSeg establishes new state-of-the-art results in performance and generalization on three segmentation tasks.
arXiv Detail & Related papers (2023-03-30T08:42:49Z) - Learning Context-aware Classifier for Semantic Segmentation [88.88198210948426]
In this paper, contextual hints are exploited via learning a context-aware classifier.
Our method is model-agnostic and can be easily applied to generic segmentation models.
With only negligible additional parameters and +2% inference time, decent performance gain has been achieved on both small and large models.
arXiv Detail & Related papers (2023-03-21T07:00:35Z) - Mask Matching Transformer for Few-Shot Segmentation [71.32725963630837]
Mask Matching Transformer (MM-Former) is a new paradigm for the few-shot segmentation task.
First, the MM-Former follows the paradigm of decompose first and then blend, allowing our method to benefit from the advanced potential objects segmenter.
We conduct extensive experiments on the popular COCO-$20i$ and Pascal-$5i$ benchmarks.
arXiv Detail & Related papers (2022-12-05T11:00:32Z) - ISIM: Iterative Self-Improved Model for Weakly Supervised Segmentation [0.34265828682659694]
Weakly Supervised Semantic Conditional (WSSS) is a challenging task aiming to learn the segmentation labels from class-level labels.
We propose a framework that employs an iterative approach in a modified encoder-decoder-based segmentation model.
Experiments performed with DeepLabv3 and UNet models show a significant gain on the Pascal VOC12 dataset.
arXiv Detail & Related papers (2022-11-22T18:14:06Z) - Task-Adaptive Feature Transformer with Semantic Enrichment for Few-Shot
Segmentation [21.276981570672064]
Few-shot learning allows machines to classify novel classes using only a few labeled samples.
We propose a learnable module that can be placed on top of existing segmentation networks for performing few-shot segmentation.
Experiments on PASCAL-$5i$ and COCO-$20i$ datasets confirm that the added modules successfully extend the capability of existing segmentators.
arXiv Detail & Related papers (2022-02-14T06:16:26Z) - CFNet: Learning Correlation Functions for One-Stage Panoptic
Segmentation [46.252118473248316]
We propose to first predict semantic-level and instance-level correlations among different locations that are utilized to enhance the backbone features.
We then feed the improved discriminative features into the corresponding segmentation heads, respectively.
We achieve state-of-the-art performance on MS with $45.1$% PQ and ADE20k with $32.6$% PQ.
arXiv Detail & Related papers (2022-01-13T05:31:14Z) - 3D Meta-Segmentation Neural Network [12.048487830494107]
We present a novel meta-learning strategy that regards the 3D shape segmentation function as a task.
By training over a number of 3D part segmentation tasks, our method is capable to learn the prior over the respective 3D segmentation function space.
We demonstrate that our model achieves superior part segmentation performance with the few-shot setting on the widely used dataset: ShapeNet.
arXiv Detail & Related papers (2021-10-08T01:47:54Z) - Segmenter: Transformer for Semantic Segmentation [79.9887988699159]
We introduce Segmenter, a transformer model for semantic segmentation.
We build on the recent Vision Transformer (ViT) and extend it to semantic segmentation.
It outperforms the state of the art on the challenging ADE20K dataset and performs on-par on Pascal Context and Cityscapes.
arXiv Detail & Related papers (2021-05-12T13:01:44Z) - SCNet: Enhancing Few-Shot Semantic Segmentation by Self-Contrastive
Background Prototypes [56.387647750094466]
Few-shot semantic segmentation aims to segment novel-class objects in a query image with only a few annotated examples.
Most of advanced solutions exploit a metric learning framework that performs segmentation through matching each pixel to a learned foreground prototype.
This framework suffers from biased classification due to incomplete construction of sample pairs with the foreground prototype only.
arXiv Detail & Related papers (2021-04-19T11:21:47Z) - Few-shot Sequence Learning with Transformers [79.87875859408955]
Few-shot algorithms aim at learning new tasks provided only a handful of training examples.
In this work we investigate few-shot learning in the setting where the data points are sequences of tokens.
We propose an efficient learning algorithm based on Transformers.
arXiv Detail & Related papers (2020-12-17T12:30:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.