Boundary-preserving Mask R-CNN
- URL: http://arxiv.org/abs/2007.08921v1
- Date: Fri, 17 Jul 2020 11:54:02 GMT
- Title: Boundary-preserving Mask R-CNN
- Authors: Tianheng Cheng and Xinggang Wang and Lichao Huang and Wenyu Liu
- Abstract summary: We propose a conceptually simple yet effective Boundary-preserving Mask R-CNN (BMask R-CNN) to leverage object boundary information to improve mask localization accuracy.
BMask R-CNN contains a boundary-preserving mask head in which object boundary and mask are mutually learned via feature fusion blocks.
Without bells and whistles, BMask R-CNN outperforms Mask R-CNN by a considerable margin on the COCO dataset.
- Score: 38.15409855290749
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Tremendous efforts have been made to improve mask localization accuracy in
instance segmentation. Modern instance segmentation methods relying on fully
convolutional networks perform pixel-wise classification, which ignores object
boundaries and shapes, leading coarse and indistinct mask prediction results
and imprecise localization. To remedy these problems, we propose a conceptually
simple yet effective Boundary-preserving Mask R-CNN (BMask R-CNN) to leverage
object boundary information to improve mask localization accuracy. BMask R-CNN
contains a boundary-preserving mask head in which object boundary and mask are
mutually learned via feature fusion blocks. As a result, the predicted masks
are better aligned with object boundaries. Without bells and whistles, BMask
R-CNN outperforms Mask R-CNN by a considerable margin on the COCO dataset; in
the Cityscapes dataset, there are more accurate boundary groundtruths
available, so that BMask R-CNN obtains remarkable improvements over Mask R-CNN.
Besides, it is not surprising to observe that BMask R-CNN obtains more obvious
improvement when the evaluation criterion requires better localization (e.g.,
AP$_{75}$) as shown in Fig.1. Code and models are available at
\url{https://github.com/hustvl/BMaskR-CNN}.
Related papers
- ColorMAE: Exploring data-independent masking strategies in Masked AutoEncoders [53.3185750528969]
Masked AutoEncoders (MAE) have emerged as a robust self-supervised framework.
We introduce a data-independent method, termed ColorMAE, which generates different binary mask patterns by filtering random noise.
We demonstrate our strategy's superiority in downstream tasks compared to random masking.
arXiv Detail & Related papers (2024-07-17T22:04:00Z) - Towards Improved Input Masking for Convolutional Neural Networks [66.99060157800403]
We propose a new masking method for CNNs we call layer masking.
We show that our method is able to eliminate or minimize the influence of the mask shape or color on the output of the model.
We also demonstrate how the shape of the mask may leak information about the class, thus affecting estimates of model reliance on class-relevant features.
arXiv Detail & Related papers (2022-11-26T19:31:49Z) - SODAR: Segmenting Objects by DynamicallyAggregating Neighboring Mask
Representations [90.8752454643737]
Recent state-of-the-art one-stage instance segmentation model SOLO divides the input image into a grid and directly predicts per grid cell object masks with fully-convolutional networks.
We observe SOLO generates similar masks for an object at nearby grid cells, and these neighboring predictions can complement each other as some may better segment certain object part.
Motivated by the observed gap, we develop a novel learning-based aggregation method that improves upon SOLO by leveraging the rich neighboring information.
arXiv Detail & Related papers (2022-02-15T13:53:03Z) - supervised adptive threshold network for instance segmentation [4.347876036795798]
Mask R-CNN method based on adaptive threshold.
layered adaptive network structure.
adaptive feature pool.
Experiments on benchmark data sets indicate that the effectiveness of the proposed model.
arXiv Detail & Related papers (2021-06-07T09:25:44Z) - Image Inpainting by End-to-End Cascaded Refinement with Mask Awareness [66.55719330810547]
Inpainting arbitrary missing regions is challenging because learning valid features for various masked regions is nontrivial.
We propose a novel mask-aware inpainting solution that learns multi-scale features for missing regions in the encoding phase.
Our framework is validated both quantitatively and qualitatively via extensive experiments on three public datasets.
arXiv Detail & Related papers (2021-04-28T13:17:47Z) - Attribution Mask: Filtering Out Irrelevant Features By Recursively
Focusing Attention on Inputs of DNNs [13.960152426268769]
Attribution methods calculate attributions that visually explain the predictions of deep neural networks (DNNs) by highlighting important parts of the input features.
In this study, we use the attributions that filter out irrelevant parts of the input features and then verify the effectiveness of this approach by measuring the classification accuracy of a pre-trained DNN.
arXiv Detail & Related papers (2021-02-15T04:12:04Z) - LevelSet R-CNN: A Deep Variational Method for Instance Segmentation [79.20048372891935]
Currently, many state of the art models are based on the Mask R-CNN framework.
We propose LevelSet R-CNN, which combines the best of both worlds by obtaining powerful feature representations.
We demonstrate the effectiveness of our approach on COCO and Cityscapes datasets.
arXiv Detail & Related papers (2020-07-30T17:52:18Z) - FibeR-CNN: Expanding Mask R-CNN to Improve Image-Based Fiber Analysis [0.0]
We propose the use of region-based convolutional neural networks (R-CNNs) to automate image-based analysis of fibers.
FibeR-CNN is able to surpass the mean average precision of Mask R-CNN by 33 % on a novel test data set of fiber images.
arXiv Detail & Related papers (2020-06-08T13:03:09Z) - Fully Convolutional Networks for Automatically Generating Image Masks to
Train Mask R-CNN [4.901462756978097]
The Mask R-CNN method achieves the best results in object detection until now, however, it is very time-consuming and laborious to get the object Masks for training.
This paper proposes a novel automatically generating image masks method for the state-of-the-art Mask R-CNN deep learning method.
Our proposed method can obtain the image masks automatically to train Mask R-CNN, and it can achieve very high classification accuracy with an over 90% mean of average precision (mAP) for segmentation.
arXiv Detail & Related papers (2020-03-03T08:09:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.