SegRefiner: Towards Model-Agnostic Segmentation Refinement with Discrete
Diffusion Process
- URL: http://arxiv.org/abs/2312.12425v1
- Date: Tue, 19 Dec 2023 18:53:47 GMT
- Title: SegRefiner: Towards Model-Agnostic Segmentation Refinement with Discrete
Diffusion Process
- Authors: Mengyu Wang, Henghui Ding, Jun Hao Liew, Jiajun Liu, Yao Zhao and
Yunchao Wei
- Abstract summary: We propose a model-agnostic solution called SegRefiner to enhance the quality of object masks produced by different segmentation models.
SegRefiner takes coarse masks as inputs and refines them using a discrete diffusion process.
It consistently improves both the segmentation metrics and boundary metrics across different types of coarse masks.
- Score: 102.18226145874007
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we explore a principal way to enhance the quality of object
masks produced by different segmentation models. We propose a model-agnostic
solution called SegRefiner, which offers a novel perspective on this problem by
interpreting segmentation refinement as a data generation process. As a result,
the refinement process can be smoothly implemented through a series of
denoising diffusion steps. Specifically, SegRefiner takes coarse masks as
inputs and refines them using a discrete diffusion process. By predicting the
label and corresponding states-transition probabilities for each pixel,
SegRefiner progressively refines the noisy masks in a conditional denoising
manner. To assess the effectiveness of SegRefiner, we conduct comprehensive
experiments on various segmentation tasks, including semantic segmentation,
instance segmentation, and dichotomous image segmentation. The results
demonstrate the superiority of our SegRefiner from multiple aspects. Firstly,
it consistently improves both the segmentation metrics and boundary metrics
across different types of coarse masks. Secondly, it outperforms previous
model-agnostic refinement methods by a significant margin. Lastly, it exhibits
a strong capability to capture extremely fine details when refining
high-resolution images. The source code and trained models are available at
https://github.com/MengyuWang826/SegRefiner.
Related papers
- SMITE: Segment Me In TimE [35.56475607621353]
We show how to segment an object in a video by employing a pre-trained text to image diffusion model and an additional tracking mechanism.
We demonstrate that our approach can effectively manage various segmentation scenarios and outperforms state-of-the-art alternatives.
arXiv Detail & Related papers (2024-10-24T08:38:20Z) - Morphologically-Aware Consensus Computation via Heuristics-based
IterATive Optimization (MACCHIatO) [1.8749305679160362]
We propose a new method to construct a binary or a probabilistic consensus segmentation based on the Fr'echet means of carefully chosen distances.
We show that it leads to binary consensus masks of intermediate size between Majority Voting and STAPLE and to different posterior probabilities than Mask Averaging and STAPLE methods.
arXiv Detail & Related papers (2023-09-14T23:28:58Z) - Synthetic Instance Segmentation from Semantic Image Segmentation Masks [15.477053085267404]
We propose a novel paradigm called Synthetic Instance (SISeg)
SISeg instance segmentation results by leveraging image masks generated by existing semantic segmentation models.
In other words, the proposed model does not need extra manpower or higher computational expenses.
arXiv Detail & Related papers (2023-08-02T05:13:02Z) - DFormer: Diffusion-guided Transformer for Universal Image Segmentation [86.73405604947459]
The proposed DFormer views universal image segmentation task as a denoising process using a diffusion model.
At inference, our DFormer directly predicts the masks and corresponding categories from a set of randomly-generated masks.
Our DFormer outperforms the recent diffusion-based panoptic segmentation method Pix2Seq-D with a gain of 3.6% on MS COCO val 2017 set.
arXiv Detail & Related papers (2023-06-06T06:33:32Z) - Denoising Diffusion Semantic Segmentation with Mask Prior Modeling [61.73352242029671]
We propose to ameliorate the semantic segmentation quality of existing discriminative approaches with a mask prior modeled by a denoising diffusion generative model.
We evaluate the proposed prior modeling with several off-the-shelf segmentors, and our experimental results on ADE20K and Cityscapes demonstrate that our approach could achieve competitively quantitative performance.
arXiv Detail & Related papers (2023-06-02T17:47:01Z) - Diffusion Models for Implicit Image Segmentation Ensembles [1.444701913511243]
We present a novel semantic segmentation method based on diffusion models.
By modifying the training and sampling scheme, we show that diffusion models can perform lesion segmentation of medical images.
Compared to state-of-the-art segmentation models, our approach yields good segmentation results and, additionally, meaningful uncertainty maps.
arXiv Detail & Related papers (2021-12-06T16:28:15Z) - SOLO: A Simple Framework for Instance Segmentation [84.00519148562606]
"instance categories" assigns categories to each pixel within an instance according to the instance's location.
"SOLO" is a simple, direct, and fast framework for instance segmentation with strong performance.
Our approach achieves state-of-the-art results for instance segmentation in terms of both speed and accuracy.
arXiv Detail & Related papers (2021-06-30T09:56:54Z) - BoundarySqueeze: Image Segmentation as Boundary Squeezing [104.43159799559464]
We propose a novel method for fine-grained high-quality image segmentation of both objects and scenes.
Inspired by dilation and erosion from morphological image processing techniques, we treat the pixel level segmentation problems as squeezing object boundary.
Our method yields large gains on COCO, Cityscapes, for both instance and semantic segmentation and outperforms previous state-of-the-art PointRend in both accuracy and speed under the same setting.
arXiv Detail & Related papers (2021-05-25T04:58:51Z) - Segmenter: Transformer for Semantic Segmentation [79.9887988699159]
We introduce Segmenter, a transformer model for semantic segmentation.
We build on the recent Vision Transformer (ViT) and extend it to semantic segmentation.
It outperforms the state of the art on the challenging ADE20K dataset and performs on-par on Pascal Context and Cityscapes.
arXiv Detail & Related papers (2021-05-12T13:01:44Z) - Post-DAE: Anatomically Plausible Segmentation via Post-Processing with
Denoising Autoencoders [19.361024564220454]
Post-DAE is a post-processing method based on denoising autoencoders (DAE)
We show how erroneous and noisy segmentation masks can be improved using Post-DAE.
arXiv Detail & Related papers (2020-06-24T15:05:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.