Weakly-supervised Instance Segmentation via Class-agnostic Learning with
Salient Images
- URL: http://arxiv.org/abs/2104.01526v1
- Date: Sun, 4 Apr 2021 03:01:52 GMT
- Title: Weakly-supervised Instance Segmentation via Class-agnostic Learning with
Salient Images
- Authors: Xinggang Wang and Jiapei Feng and Bin Hu and Qi Ding and Longjin Ran
and Xiaoxin Chen and Wenyu Liu
- Abstract summary: We propose a box-supervised class-agnostic object segmentation (BoxCaseg) based solution for weakly-supervised instance segmentation.
The BoxCaseg model is jointly trained using box-supervised images and salient images in a multi-task learning manner.
The weakly-supervised Mask R-CNN is on par with fully-supervised Mask R-CNN on PASCAL VOC and significantly outperforms previous state-of-the-art box-supervised instance segmentation methods on COCO.
- Score: 34.498542318157284
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Humans have a strong class-agnostic object segmentation ability and can
outline boundaries of unknown objects precisely, which motivates us to propose
a box-supervised class-agnostic object segmentation (BoxCaseg) based solution
for weakly-supervised instance segmentation. The BoxCaseg model is jointly
trained using box-supervised images and salient images in a multi-task learning
manner. The fine-annotated salient images provide class-agnostic and precise
object localization guidance for box-supervised images. The object masks
predicted by a pretrained BoxCaseg model are refined via a novel merged and
dropped strategy as proxy ground truth to train a Mask R-CNN for
weakly-supervised instance segmentation. Only using $7991$ salient images, the
weakly-supervised Mask R-CNN is on par with fully-supervised Mask R-CNN on
PASCAL VOC and significantly outperforms previous state-of-the-art
box-supervised instance segmentation methods on COCO. The source code,
pretrained models and datasets are available at
\url{https://github.com/hustvl/BoxCaseg}.
Related papers
- Boosting Unsupervised Semantic Segmentation with Principal Mask Proposals [15.258631373740686]
Unsupervised semantic segmentation aims to automatically partition images into semantically meaningful regions by identifying global semantic categories within an image corpus without any form of annotation.
We present PriMaPs - Principal Mask Proposals - decomposing images into semantically meaningful masks based on their feature representation.
This allows us to realize unsupervised semantic segmentation by fitting class prototypes to PriMaPs with a expectation-maximization algorithm, PriMaPs-EM.
arXiv Detail & Related papers (2024-04-25T17:58:09Z) - Mask-free OVIS: Open-Vocabulary Instance Segmentation without Manual
Mask Annotations [86.47908754383198]
Open-Vocabulary (OV) methods leverage large-scale image-caption pairs and vision-language models to learn novel categories.
Our method generates pseudo-mask annotations by leveraging the localization ability of a pre-trained vision-language model for objects present in image-caption pairs.
Our method trained with just pseudo-masks significantly improves the mAP scores on the MS-COCO dataset and OpenImages dataset.
arXiv Detail & Related papers (2023-03-29T17:58:39Z) - Learning Open-vocabulary Semantic Segmentation Models From Natural
Language Supervision [49.905448429974804]
We consider the problem of open-vocabulary semantic segmentation (OVS), which aims to segment objects of arbitrary classes instead of pre-defined, closed-set categories.
We propose a transformer-based model for OVS, termed as OVSegmentor, which exploits web-crawled image-text pairs for pre-training.
Our model achieves superior segmentation results over the state-of-the-art method by using only 3% data (4M vs 134M) for pre-training.
arXiv Detail & Related papers (2023-01-22T13:10:05Z) - Location-Aware Self-Supervised Transformers [74.76585889813207]
We propose to pretrain networks for semantic segmentation by predicting the relative location of image parts.
We control the difficulty of the task by masking a subset of the reference patch features visible to those of the query.
Our experiments show that this location-aware pretraining leads to representations that transfer competitively to several challenging semantic segmentation benchmarks.
arXiv Detail & Related papers (2022-12-05T16:24:29Z) - Exploiting Shape Cues for Weakly Supervised Semantic Segmentation [15.791415215216029]
Weakly supervised semantic segmentation (WSSS) aims to produce pixel-wise class predictions with only image-level labels for training.
We propose to exploit shape information to supplement the texture-biased property of convolutional neural networks (CNNs)
We further refine the predictions in an online fashion with a novel refinement method that takes into account both the class and the color affinities.
arXiv Detail & Related papers (2022-08-08T17:25:31Z) - GANSeg: Learning to Segment by Unsupervised Hierarchical Image
Generation [16.900404701997502]
We propose a GAN-based approach that generates images conditioned on latent masks.
We show that such mask-conditioned image generation can be learned faithfully when conditioning the masks in a hierarchical manner.
It also lets us generate image-mask pairs for training a segmentation network, which outperforms the state-of-the-art unsupervised segmentation methods on established benchmarks.
arXiv Detail & Related papers (2021-12-02T07:57:56Z) - Box-Adapt: Domain-Adaptive Medical Image Segmentation using Bounding
BoxSupervision [52.45336255472669]
We propose a weakly supervised do-main adaptation setting for deep learning.
Box-Adapt fully explores the fine-grained segmenta-tion mask in the source domain and the weak bounding box in the target domain.
We demonstrate the effectiveness of our method in the liver segmentation task.
arXiv Detail & Related papers (2021-08-19T01:51:04Z) - BBAM: Bounding Box Attribution Map for Weakly Supervised Semantic and
Instance Segmentation [19.55647093153416]
Weakly supervised segmentation methods using bounding box annotations focus on obtaining a pixel-level mask from each box containing an object.
In this work, we utilize higher-level information from the behavior of a trained object detector, by seeking the smallest areas of the image from which the object detector produces almost the same result as it does from the whole image.
These areas constitute a bounding-box attribution map (BBAM), which identifies the target object in its bounding box and thus serves as pseudo ground-truth for weakly supervised semantic and COCO instance segmentation.
arXiv Detail & Related papers (2021-03-16T08:29:33Z) - Learning Class-Agnostic Pseudo Mask Generation for Box-Supervised
Semantic Segmentation [156.9155100983315]
We seek for a more accurate learning-based class-agnostic pseudo mask generator tailored to box-supervised semantic segmentation.
Our method can further close the performance gap between box-supervised and fully-supervised models.
arXiv Detail & Related papers (2021-03-09T14:54:54Z) - Instance Localization for Self-supervised Detection Pretraining [68.24102560821623]
We propose a new self-supervised pretext task, called instance localization.
We show that integration of bounding boxes into pretraining promotes better task alignment and architecture alignment for transfer learning.
Experimental results demonstrate that our approach yields state-of-the-art transfer learning results for object detection.
arXiv Detail & Related papers (2021-02-16T17:58:57Z) - Prior to Segment: Foreground Cues for Weakly Annotated Classes in
Partially Supervised Instance Segmentation [3.192503074844774]
Partially supervised instance segmentation aims to improve mask prediction with limited mask labels by utilizing the more abundant weak box labels.
We show that a class agnostic mask head, commonly used in partially supervised instance segmentation, has difficulties learning a general concept of foreground for the weakly annotated classes.
We introduce an object mask prior (OMP) that provides the mask head with the general concept of foreground implicitly learned by the box classification head under the supervision of all classes.
arXiv Detail & Related papers (2020-11-23T23:15:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.