Unsupervised Part Discovery by Unsupervised Disentanglement
- URL: http://arxiv.org/abs/2009.04264v2
- Date: Thu, 10 Sep 2020 09:31:48 GMT
- Title: Unsupervised Part Discovery by Unsupervised Disentanglement
- Authors: Sandro Braun, Patrick Esser, Bj\"orn Ommer
- Abstract summary: Part segmentations provide information about part localizations on the level of individual pixels.
Large annotation costs limit the scalability of supervised algorithms to other object categories.
Our work demonstrates the feasibility to discover semantic part segmentations without supervision.
- Score: 10.664434993386525
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We address the problem of discovering part segmentations of articulated
objects without supervision. In contrast to keypoints, part segmentations
provide information about part localizations on the level of individual pixels.
Capturing both locations and semantics, they are an attractive target for
supervised learning approaches. However, large annotation costs limit the
scalability of supervised algorithms to other object categories than humans.
Unsupervised approaches potentially allow to use much more data at a lower
cost. Most existing unsupervised approaches focus on learning abstract
representations to be refined with supervision into the final representation.
Our approach leverages a generative model consisting of two disentangled
representations for an object's shape and appearance and a latent variable for
the part segmentation. From a single image, the trained model infers a semantic
part segmentation map. In experiments, we compare our approach to previous
state-of-the-art approaches and observe significant gains in segmentation
accuracy and shape consistency. Our work demonstrates the feasibility to
discover semantic part segmentations without supervision.
Related papers
- Weakly-supervised Contrastive Learning for Unsupervised Object Discovery [52.696041556640516]
Unsupervised object discovery is promising due to its ability to discover objects in a generic manner.
We design a semantic-guided self-supervised learning model to extract high-level semantic features from images.
We introduce Principal Component Analysis (PCA) to localize object regions.
arXiv Detail & Related papers (2023-07-07T04:03:48Z) - Exploring Open-Vocabulary Semantic Segmentation without Human Labels [76.15862573035565]
We present ZeroSeg, a novel method that leverages the existing pretrained vision-language model (VL) to train semantic segmentation models.
ZeroSeg overcomes this by distilling the visual concepts learned by VL models into a set of segment tokens, each summarizing a localized region of the target image.
Our approach achieves state-of-the-art performance when compared to other zero-shot segmentation methods under the same training data.
arXiv Detail & Related papers (2023-06-01T08:47:06Z) - Towards Open-World Segmentation of Parts [16.056921233445784]
We propose to explore a class-agnostic part segmentation task.
We argue that models trained without part classes can better localize parts and segment them on objects unseen in training.
We show notable and consistent gains by our approach, essentially a critical step towards open-world part segmentation.
arXiv Detail & Related papers (2023-05-26T10:34:58Z) - Self-Supervised Learning of Object Parts for Semantic Segmentation [7.99536002595393]
We argue that self-supervised learning of object parts is a solution to this issue.
Our method surpasses the state-of-the-art on three semantic segmentation benchmarks by 17%-3%.
arXiv Detail & Related papers (2022-04-27T17:55:17Z) - Unsupervised Part Discovery from Contrastive Reconstruction [90.88501867321573]
The goal of self-supervised visual representation learning is to learn strong, transferable image representations.
We propose an unsupervised approach to object part discovery and segmentation.
Our method yields semantic parts consistent across fine-grained but visually distinct categories.
arXiv Detail & Related papers (2021-11-11T17:59:42Z) - Revisiting Contrastive Methods for Unsupervised Learning of Visual
Representations [78.12377360145078]
Contrastive self-supervised learning has outperformed supervised pretraining on many downstream tasks like segmentation and object detection.
In this paper, we first study how biases in the dataset affect existing methods.
We show that current contrastive approaches work surprisingly well across: (i) object- versus scene-centric, (ii) uniform versus long-tailed and (iii) general versus domain-specific datasets.
arXiv Detail & Related papers (2021-06-10T17:59:13Z) - Unsupervised Co-part Segmentation through Assembly [42.874278526843305]
We propose an unsupervised learning approach for co-part segmentation from images.
We leverage motion information embedded in videos and explicitly extract latent representations to segment meaningful object parts.
We show that our approach can achieve meaningful and compact part segmentation, outperforming state-of-the-art approaches on diverse benchmarks.
arXiv Detail & Related papers (2021-06-10T16:22:53Z) - Unsupervised Part Segmentation through Disentangling Appearance and
Shape [37.206922180245265]
We study the problem of unsupervised discovery and segmentation of object parts.
Recent unsupervised methods have greatly relaxed the dependency on annotated data.
We develop a novel approach by disentangling the appearance and shape representations of object parts.
arXiv Detail & Related papers (2021-05-26T08:59:31Z) - Unsupervised Semantic Segmentation by Contrasting Object Mask Proposals [78.12377360145078]
We introduce a novel two-step framework that adopts a predetermined prior in a contrastive optimization objective to learn pixel embeddings.
This marks a large deviation from existing works that relied on proxy tasks or end-to-end clustering.
In particular, when fine-tuning the learned representations using just 1% of labeled examples on PASCAL, we outperform supervised ImageNet pre-training by 7.1% mIoU.
arXiv Detail & Related papers (2021-02-11T18:54:47Z) - Global-Local Bidirectional Reasoning for Unsupervised Representation
Learning of 3D Point Clouds [109.0016923028653]
We learn point cloud representation by bidirectional reasoning between the local structures and the global shape without human supervision.
We show that our unsupervised model surpasses the state-of-the-art supervised methods on both synthetic and real-world 3D object classification datasets.
arXiv Detail & Related papers (2020-03-29T08:26:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.