Unsupervised Part Segmentation through Disentangling Appearance and
Shape
- URL: http://arxiv.org/abs/2105.12405v1
- Date: Wed, 26 May 2021 08:59:31 GMT
- Title: Unsupervised Part Segmentation through Disentangling Appearance and
Shape
- Authors: Shilong Liu, Lei Zhang, Xiao Yang, Hang Su, Jun Zhu
- Abstract summary: We study the problem of unsupervised discovery and segmentation of object parts.
Recent unsupervised methods have greatly relaxed the dependency on annotated data.
We develop a novel approach by disentangling the appearance and shape representations of object parts.
- Score: 37.206922180245265
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We study the problem of unsupervised discovery and segmentation of object
parts, which, as an intermediate local representation, are capable of finding
intrinsic object structure and providing more explainable recognition results.
Recent unsupervised methods have greatly relaxed the dependency on annotated
data which are costly to obtain, but still rely on additional information such
as object segmentation mask or saliency map. To remove such a dependency and
further improve the part segmentation performance, we develop a novel approach
by disentangling the appearance and shape representations of object parts
followed with reconstruction losses without using additional object mask
information. To avoid degenerated solutions, a bottleneck block is designed to
squeeze and expand the appearance representation, leading to a more effective
disentanglement between geometry and appearance. Combined with a
self-supervised part classification loss and an improved geometry concentration
constraint, we can segment more consistent parts with semantic meanings.
Comprehensive experiments on a wide variety of objects such as face, bird, and
PASCAL VOC objects demonstrate the effectiveness of the proposed method.
Related papers
- EAGLE: Eigen Aggregation Learning for Object-Centric Unsupervised Semantic Segmentation [5.476136494434766]
We introduce EiCue, a technique providing semantic and structural cues through an eigenbasis derived from semantic similarity matrix.
We guide our model to learn object-level representations with intra- and inter-image object-feature consistency.
Experiments on COCO-Stuff, Cityscapes, and Potsdam-3 datasets demonstrate the state-of-the-art USS results.
arXiv Detail & Related papers (2024-03-03T11:24:16Z) - Weakly-supervised Contrastive Learning for Unsupervised Object Discovery [52.696041556640516]
Unsupervised object discovery is promising due to its ability to discover objects in a generic manner.
We design a semantic-guided self-supervised learning model to extract high-level semantic features from images.
We introduce Principal Component Analysis (PCA) to localize object regions.
arXiv Detail & Related papers (2023-07-07T04:03:48Z) - Deep Spectral Methods: A Surprisingly Strong Baseline for Unsupervised
Semantic Segmentation and Localization [98.46318529630109]
We take inspiration from traditional spectral segmentation methods by reframing image decomposition as a graph partitioning problem.
We find that these eigenvectors already decompose an image into meaningful segments, and can be readily used to localize objects in a scene.
By clustering the features associated with these segments across a dataset, we can obtain well-delineated, nameable regions.
arXiv Detail & Related papers (2022-05-16T17:47:44Z) - Self-Supervised Video Object Segmentation via Cutout Prediction and
Tagging [117.73967303377381]
We propose a novel self-supervised Video Object (VOS) approach that strives to achieve better object-background discriminability.
Our approach is based on a discriminative learning loss formulation that takes into account both object and background information.
Our proposed approach, CT-VOS, achieves state-of-the-art results on two challenging benchmarks: DAVIS-2017 and Youtube-VOS.
arXiv Detail & Related papers (2022-04-22T17:53:27Z) - Unsupervised Part Discovery from Contrastive Reconstruction [90.88501867321573]
The goal of self-supervised visual representation learning is to learn strong, transferable image representations.
We propose an unsupervised approach to object part discovery and segmentation.
Our method yields semantic parts consistent across fine-grained but visually distinct categories.
arXiv Detail & Related papers (2021-11-11T17:59:42Z) - Unsupervised Discovery of the Long-Tail in Instance Segmentation Using
Hierarchical Self-Supervision [3.841232411073827]
We propose a method that can perform unsupervised discovery of long-tail categories in instance segmentation.
Our model is able to discover novel and more fine-grained objects than the common categories.
We show that the model achieves competitive quantitative results on LVIS as compared to the supervised and partially supervised methods.
arXiv Detail & Related papers (2021-04-02T22:05:03Z) - Self-supervised Segmentation via Background Inpainting [96.10971980098196]
We introduce a self-supervised detection and segmentation approach that can work with single images captured by a potentially moving camera.
We exploit a self-supervised loss function that we exploit to train a proposal-based segmentation network.
We apply our method to human detection and segmentation in images that visually depart from those of standard benchmarks and outperform existing self-supervised methods.
arXiv Detail & Related papers (2020-11-11T08:34:40Z) - Unsupervised Part Discovery by Unsupervised Disentanglement [10.664434993386525]
Part segmentations provide information about part localizations on the level of individual pixels.
Large annotation costs limit the scalability of supervised algorithms to other object categories.
Our work demonstrates the feasibility to discover semantic part segmentations without supervision.
arXiv Detail & Related papers (2020-09-09T12:34:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.