Boundary Knowledge Translation based Reference Semantic Segmentation
- URL: http://arxiv.org/abs/2108.01075v1
- Date: Sun, 1 Aug 2021 07:40:09 GMT
- Title: Boundary Knowledge Translation based Reference Semantic Segmentation
- Authors: Lechao Cheng, Zunlei Feng, Xinchao Wang, Ya Jie Liu, Jie Lei, Mingli
Song
- Abstract summary: We introduce a Reference Reference segmentation Network (Ref-Net) to conduct visual boundary knowledge translation.
Inspired by the human recognition mechanism, RSMTM is devised only to segment the same category objects based on the features of the reference objects.
With tens of finely-grained annotated samples as guidance, Ref-Net achieves results on par with fully supervised methods on six datasets.
- Score: 62.60078935335371
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Given a reference object of an unknown type in an image, human observers can
effortlessly find the objects of the same category in another image and
precisely tell their visual boundaries. Such visual cognition capability of
humans seems absent from the current research spectrum of computer vision.
Existing segmentation networks, for example, rely on a humongous amount of
labeled data, which is laborious and costly to collect and annotate; besides,
the performance of segmentation networks tend to downgrade as the number of the
category increases. In this paper, we introduce a novel Reference semantic
segmentation Network (Ref-Net) to conduct visual boundary knowledge
translation. Ref-Net contains a Reference Segmentation Module (RSM) and a
Boundary Knowledge Translation Module (BKTM). Inspired by the human recognition
mechanism, RSM is devised only to segment the same category objects based on
the features of the reference objects. BKTM, on the other hand, introduces two
boundary discriminator branches to conduct inner and outer boundary
segmentation of the target objectin an adversarial manner, and translate the
annotated boundary knowledge of open-source datasets into the segmentation
network. Exhaustive experiments demonstrate that, with tens of finely-grained
annotated samples as guidance, Ref-Net achieves results on par with fully
supervised methods on six datasets.
Related papers
- Exploiting Object-based and Segmentation-based Semantic Features for Deep Learning-based Indoor Scene Classification [0.5572976467442564]
The work described in this paper uses both semantic information, obtained from object detection, and semantic segmentation techniques.
A novel approach that uses a semantic segmentation mask to provide Hu-moments-based segmentation categories' shape characterization, designated by Hu-Moments Features (SHMFs) is proposed.
A three-main-branch network, designated by GOS$2$F$2$App, that exploits deep-learning-based global features, object-based features, and semantic segmentation-based features is also proposed.
arXiv Detail & Related papers (2024-04-11T13:37:51Z) - SAM-Assisted Remote Sensing Imagery Semantic Segmentation with Object
and Boundary Constraints [9.238103649037951]
We present a framework aimed at leveraging the raw output of SAM by exploiting two novel concepts called SAM-Generated Object (SGO) and SAM-Generated Boundary (SGB)
Taking into account the content characteristics of SGO, we introduce the concept of object consistency to leverage segmented regions lacking semantic information.
The boundary loss capitalizes on the distinctive features of SGB by directing the model's attention to the boundary information of the object.
arXiv Detail & Related papers (2023-12-05T03:33:47Z) - Lidar Panoptic Segmentation and Tracking without Bells and Whistles [48.078270195629415]
We propose a detection-centric network for lidar segmentation and tracking.
One of the core components of our network is the object instance detection branch.
We evaluate our method on several 3D/4D LPS benchmarks and observe that our model establishes a new state-of-the-art among open-sourced models.
arXiv Detail & Related papers (2023-10-19T04:44:43Z) - Weakly-supervised Contrastive Learning for Unsupervised Object Discovery [52.696041556640516]
Unsupervised object discovery is promising due to its ability to discover objects in a generic manner.
We design a semantic-guided self-supervised learning model to extract high-level semantic features from images.
We introduce Principal Component Analysis (PCA) to localize object regions.
arXiv Detail & Related papers (2023-07-07T04:03:48Z) - Open-world Semantic Segmentation via Contrasting and Clustering
Vision-Language Embedding [95.78002228538841]
We propose a new open-world semantic segmentation pipeline that makes the first attempt to learn to segment semantic objects of various open-world categories without any efforts on dense annotations.
Our method can directly segment objects of arbitrary categories, outperforming zero-shot segmentation methods that require data labeling on three benchmark datasets.
arXiv Detail & Related papers (2022-07-18T09:20:04Z) - Visual Boundary Knowledge Translation for Foreground Segmentation [57.32522585756404]
We make an attempt towards building models that explicitly account for visual boundary knowledge, in hope to reduce the training effort on segmenting unseen categories.
With only tens of labeled samples as guidance, Trans-Net achieves close results on par with fully supervised methods.
arXiv Detail & Related papers (2021-08-01T07:10:25Z) - Bidirectional Graph Reasoning Network for Panoptic Segmentation [126.06251745669107]
We introduce a Bidirectional Graph Reasoning Network (BGRNet) to mine the intra-modular and intermodular relations within and between foreground things and background stuff classes.
BGRNet first constructs image-specific graphs in both instance and semantic segmentation branches that enable flexible reasoning at the proposal level and class level.
arXiv Detail & Related papers (2020-04-14T02:32:10Z) - Universal-RCNN: Universal Object Detector via Transferable Graph R-CNN [117.80737222754306]
We present a novel universal object detector called Universal-RCNN.
We first generate a global semantic pool by integrating all high-level semantic representation of all the categories.
An Intra-Domain Reasoning Module learns and propagates the sparse graph representation within one dataset guided by a spatial-aware GCN.
arXiv Detail & Related papers (2020-02-18T07:57:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.