Semantically Meaningful Class Prototype Learning for One-Shot Image
Semantic Segmentation
- URL: http://arxiv.org/abs/2102.10935v1
- Date: Mon, 22 Feb 2021 12:07:35 GMT
- Title: Semantically Meaningful Class Prototype Learning for One-Shot Image
Semantic Segmentation
- Authors: Tao Chen, Guosen Xie, Yazhou Yao, Qiong Wang, Fumin Shen, Zhenmin
Tang, and Jian Zhang
- Abstract summary: One-shot semantic image segmentation aims to segment the object regions for the novel class with only one annotated image.
Recent works adopt the episodic training strategy to mimic the expected situation at testing time.
We propose to leverage the multi-class label information during the episodic training. It will encourage the network to generate more semantically meaningful features for each category.
- Score: 58.96902899546075
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: One-shot semantic image segmentation aims to segment the object regions for
the novel class with only one annotated image. Recent works adopt the episodic
training strategy to mimic the expected situation at testing time. However,
these existing approaches simulate the test conditions too strictly during the
training process, and thus cannot make full use of the given label information.
Besides, these approaches mainly focus on the foreground-background target
class segmentation setting. They only utilize binary mask labels for training.
In this paper, we propose to leverage the multi-class label information during
the episodic training. It will encourage the network to generate more
semantically meaningful features for each category. After integrating the
target class cues into the query features, we then propose a pyramid feature
fusion module to mine the fused features for the final classifier. Furthermore,
to take more advantage of the support image-mask pair, we propose a
self-prototype guidance branch to support image segmentation. It can constrain
the network for generating more compact features and a robust prototype for
each semantic class. For inference, we propose a fused prototype guidance
branch for the segmentation of the query image. Specifically, we leverage the
prediction of the query image to extract the pseudo-prototype and combine it
with the initial prototype. Then we utilize the fused prototype to guide the
final segmentation of the query image. Extensive experiments demonstrate the
superiority of our proposed approach.
Related papers
- IFSENet : Harnessing Sparse Iterations for Interactive Few-shot Segmentation Excellence [2.822194296769473]
Few-shot segmentation techniques reduce the required number of images to learn to segment a new class.
interactive segmentation techniques only focus on incrementally improving the segmentation of one object at a time.
We combine the two concepts to drastically reduce the effort required to train segmentation models for novel classes.
arXiv Detail & Related papers (2024-03-22T10:15:53Z) - Learning Semantic Segmentation with Query Points Supervision on Aerial Images [57.09251327650334]
We present a weakly supervised learning algorithm to train semantic segmentation algorithms.
Our proposed approach performs accurate semantic segmentation and improves efficiency by significantly reducing the cost and time required for manual annotation.
arXiv Detail & Related papers (2023-09-11T14:32:04Z) - Iterative Few-shot Semantic Segmentation from Image Label Text [36.53926941601841]
Few-shot semantic segmentation aims to learn to segment unseen class objects with the guidance of only a few support images.
We propose a general framework to generate coarse masks with the help of the powerful vision-language model CLIP.
Our method owns an excellent generalization ability for the images in the wild and uncommon classes.
arXiv Detail & Related papers (2023-03-10T01:48:14Z) - APANet: Adaptive Prototypes Alignment Network for Few-Shot Semantic
Segmentation [56.387647750094466]
Few-shot semantic segmentation aims to segment novel-class objects in a given query image with only a few labeled support images.
Most advanced solutions exploit a metric learning framework that performs segmentation through matching each query feature to a learned class-specific prototype.
We present an adaptive prototype representation by introducing class-specific and class-agnostic prototypes.
arXiv Detail & Related papers (2021-11-24T04:38:37Z) - SCNet: Enhancing Few-Shot Semantic Segmentation by Self-Contrastive
Background Prototypes [56.387647750094466]
Few-shot semantic segmentation aims to segment novel-class objects in a query image with only a few annotated examples.
Most of advanced solutions exploit a metric learning framework that performs segmentation through matching each pixel to a learned foreground prototype.
This framework suffers from biased classification due to incomplete construction of sample pairs with the foreground prototype only.
arXiv Detail & Related papers (2021-04-19T11:21:47Z) - Part-aware Prototype Network for Few-shot Semantic Segmentation [50.581647306020095]
We propose a novel few-shot semantic segmentation framework based on the prototype representation.
Our key idea is to decompose the holistic class representation into a set of part-aware prototypes.
We develop a novel graph neural network model to generate and enhance the proposed part-aware prototypes.
arXiv Detail & Related papers (2020-07-13T11:03:09Z) - CRNet: Cross-Reference Networks for Few-Shot Segmentation [59.85183776573642]
Few-shot segmentation aims to learn a segmentation model that can be generalized to novel classes with only a few training images.
With a cross-reference mechanism, our network can better find the co-occurrent objects in the two images.
Experiments on the PASCAL VOC 2012 dataset show that our network achieves state-of-the-art performance.
arXiv Detail & Related papers (2020-03-24T04:55:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.