Unsupervised Landmark Discovery Using Consistency Guided Bottleneck
- URL: http://arxiv.org/abs/2309.10518v1
- Date: Tue, 19 Sep 2023 10:57:53 GMT
- Title: Unsupervised Landmark Discovery Using Consistency Guided Bottleneck
- Authors: Mamona Awan, Muhammad Haris Khan, Sanoojan Baliah, Muhammad Ahmad
Waseem, Salman Khan, Fahad Shahbaz Khan and Arif Mahmood
- Abstract summary: We introduce a consistency-guided bottleneck in an image reconstruction-based pipeline.
We propose obtaining pseudo-supervision via forming landmark correspondence across images.
The consistency then modulates the uncertainty of the discovered landmarks in the generation of adaptive heatmaps.
- Score: 63.624186864522315
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We study a challenging problem of unsupervised discovery of object landmarks.
Many recent methods rely on bottlenecks to generate 2D Gaussian heatmaps
however, these are limited in generating informed heatmaps while training,
presumably due to the lack of effective structural cues. Also, it is assumed
that all predicted landmarks are semantically relevant despite having no ground
truth supervision. In the current work, we introduce a consistency-guided
bottleneck in an image reconstruction-based pipeline that leverages landmark
consistency, a measure of compatibility score with the pseudo-ground truth to
generate adaptive heatmaps. We propose obtaining pseudo-supervision via forming
landmark correspondence across images. The consistency then modulates the
uncertainty of the discovered landmarks in the generation of adaptive heatmaps
which rank consistent landmarks above their noisy counterparts, providing
effective structural information for improved robustness. Evaluations on five
diverse datasets including MAFL, AFLW, LS3D, Cats, and Shoes demonstrate
excellent performance of the proposed approach compared to the existing
state-of-the-art methods. Our code is publicly available at
https://github.com/MamonaAwan/CGB_ULD.
Related papers
- Assessing and improving reliability of neighbor embedding methods: a map-continuity perspective [2.969441406380581]
Neighbor embedding methods, such as t-SNE, UMAP, and LargeVis, are a family of popular visualization methods.
Recent studies suggest that these methods often produce visual artifacts, potentially leading to incorrect scientific conclusions.
We introduce a novel conceptual and computational framework, LOO-map, that learns the embedding maps based on a classical statistical idea.
arXiv Detail & Related papers (2024-10-22T01:40:43Z) - Disentangled Representation Learning with the Gromov-Monge Gap [65.73194652234848]
Learning disentangled representations from unlabelled data is a fundamental challenge in machine learning.
We introduce a novel approach to disentangled representation learning based on quadratic optimal transport.
We demonstrate the effectiveness of our approach for quantifying disentanglement across four standard benchmarks.
arXiv Detail & Related papers (2024-07-10T16:51:32Z) - Learning Real-World Image De-Weathering with Imperfect Supervision [57.748585821252824]
Existing real-world de-weathering datasets often exhibit inconsistent illumination, position, and textures between the ground-truth images and the input degraded images.
We develop a Consistent Label Constructor (CLC) to generate a pseudo-label as consistent as possible with the input degraded image.
We combine the original imperfect labels and pseudo-labels to jointly supervise the de-weathering model by the proposed Information Allocation Strategy.
arXiv Detail & Related papers (2023-10-23T14:02:57Z) - VIBUS: Data-efficient 3D Scene Parsing with VIewpoint Bottleneck and
Uncertainty-Spectrum Modeling [2.0624279915507047]
Training 3D scene parsing models with sparse supervision is an intriguing alternative.
We term this task as data-efficient 3D scene parsing.
We propose an effective two-stage framework named VIBUS to resolve it.
arXiv Detail & Related papers (2022-10-20T17:59:57Z) - Pretrained equivariant features improve unsupervised landmark discovery [69.02115180674885]
We formulate a two-step unsupervised approach that overcomes this challenge by first learning powerful pixel-based features.
Our method produces state-of-the-art results in several challenging landmark detection datasets.
arXiv Detail & Related papers (2021-04-07T05:42:11Z) - Deep Structured Prediction for Facial Landmark Detection [59.60946775628646]
This paper proposes a method for deep structured facial landmark detection based on combining a deep Convolutional Network with a Conditional Random Field.
We demonstrate its superior performance to existing state-of-the-art techniques in facial landmark detection.
arXiv Detail & Related papers (2020-10-18T17:09:24Z) - Gaussian Process Gradient Maps for Loop-Closure Detection in
Unstructured Planetary Environments [17.276441789710574]
The ability to recognize previously mapped locations is an essential feature for autonomous systems.
Unstructured planetary-like environments pose a major challenge to these systems due to the similarity of the terrain.
This paper presents a method to solve the loop closure problem using only spatial information.
arXiv Detail & Related papers (2020-09-01T04:41:40Z) - BRUL\`E: Barycenter-Regularized Unsupervised Landmark Extraction [2.2758845733923687]
Unsupervised retrieval of image features is vital for many computer vision tasks where the annotation is missing or scarce.
We propose a new unsupervised approach to detect the landmarks in images, validating it on the popular task of human face key-points extraction.
The method is based on the idea of auto-encoding the wanted landmarks in the latent space while discarding the non-essential information.
arXiv Detail & Related papers (2020-06-20T20:04:00Z) - Rethinking Localization Map: Towards Accurate Object Perception with
Self-Enhancement Maps [78.2581910688094]
This work introduces a novel self-enhancement method to harvest accurate object localization maps and object boundaries with only category labels as supervision.
In particular, the proposed Self-Enhancement Maps achieve the state-of-the-art localization accuracy of 54.88% on ILSVRC.
arXiv Detail & Related papers (2020-06-09T12:35:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.