Variational Clustering: Leveraging Variational Autoencoders for Image
Clustering
- URL: http://arxiv.org/abs/2005.04613v1
- Date: Sun, 10 May 2020 09:34:48 GMT
- Title: Variational Clustering: Leveraging Variational Autoencoders for Image
Clustering
- Authors: Vignesh Prasad, Dipanjan Das, Brojeshwar Bhowmick
- Abstract summary: Variational Autoencoders (VAEs) naturally lend themselves to learning data distributions in a latent space.
We propose a method based on VAEs where we use a Gaussian Mixture prior to help cluster the images accurately.
Our method simultaneously learns a prior that captures the latent distribution of the images and a posterior to help discriminate well between data points.
- Score: 8.465172258675763
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent advances in deep learning have shown their ability to learn strong
feature representations for images. The task of image clustering naturally
requires good feature representations to capture the distribution of the data
and subsequently differentiate data points from one another. Often these two
aspects are dealt with independently and thus traditional feature learning
alone does not suffice in partitioning the data meaningfully. Variational
Autoencoders (VAEs) naturally lend themselves to learning data distributions in
a latent space. Since we wish to efficiently discriminate between different
clusters in the data, we propose a method based on VAEs where we use a Gaussian
Mixture prior to help cluster the images accurately. We jointly learn the
parameters of both the prior and the posterior distributions. Our method
represents a true Gaussian Mixture VAE. This way, our method simultaneously
learns a prior that captures the latent distribution of the images and a
posterior to help discriminate well between data points. We also propose a
novel reparametrization of the latent space consisting of a mixture of discrete
and continuous variables. One key takeaway is that our method generalizes
better across different datasets without using any pre-training or learnt
models, unlike existing methods, allowing it to be trained from scratch in an
end-to-end manner. We verify our efficacy and generalizability experimentally
by achieving state-of-the-art results among unsupervised methods on a variety
of datasets. To the best of our knowledge, we are the first to pursue image
clustering using VAEs in a purely unsupervised manner on real image datasets.
Related papers
- Mitigating Bias Using Model-Agnostic Data Attribution [2.9868610316099335]
Mitigating bias in machine learning models is a critical endeavor for ensuring fairness and equity.
We propose a novel approach to address bias by leveraging pixel image attributions to identify and regularize regions of images containing bias attributes.
arXiv Detail & Related papers (2024-05-08T13:00:56Z) - SatSynth: Augmenting Image-Mask Pairs through Diffusion Models for Aerial Semantic Segmentation [69.42764583465508]
We explore the potential of generative image diffusion to address the scarcity of annotated data in earth observation tasks.
To the best of our knowledge, we are the first to generate both images and corresponding masks for satellite segmentation.
arXiv Detail & Related papers (2024-03-25T10:30:22Z) - Understanding normalization in contrastive representation learning and out-of-distribution detection [0.0]
We propose a simple method based on contrastive learning, which incorporates out-of-distribution data by discriminating against normal samples in the contrastive layer space.
Our approach can be applied flexibly as an outlier exposure (OE) approach, or as a fully self-supervised learning approach.
The high-quality features learned through contrastive learning consistently enhance performance in OE scenarios, even when the available out-of-distribution dataset is not diverse enough.
arXiv Detail & Related papers (2023-12-23T16:05:47Z) - Cluster-level pseudo-labelling for source-free cross-domain facial
expression recognition [94.56304526014875]
We propose the first Source-Free Unsupervised Domain Adaptation (SFUDA) method for Facial Expression Recognition (FER)
Our method exploits self-supervised pretraining to learn good feature representations from the target data.
We validate the effectiveness of our method in four adaptation setups, proving that it consistently outperforms existing SFUDA methods when applied to FER.
arXiv Detail & Related papers (2022-10-11T08:24:50Z) - Invariance Learning in Deep Neural Networks with Differentiable Laplace
Approximations [76.82124752950148]
We develop a convenient gradient-based method for selecting the data augmentation.
We use a differentiable Kronecker-factored Laplace approximation to the marginal likelihood as our objective.
arXiv Detail & Related papers (2022-02-22T02:51:11Z) - Deep Variational Clustering Framework for Self-labeling of Large-scale
Medical Images [2.9617653204513656]
We propose a Deep Variational Clustering (DVC) framework for unsupervised representation learning and clustering of large-scale medical images.
In this approach, the probabilistic decoder helps to prevent the distortion of data points in the latent space and to preserve the local structure of data generating distribution.
Our experimental results show that our proposed framework generalizes better across different datasets.
arXiv Detail & Related papers (2021-09-22T15:01:52Z) - Unlabeled Data Guided Semi-supervised Histopathology Image Segmentation [34.45302976822067]
Semi-supervised learning (SSL) based on generative methods has been proven to be effective in utilizing diverse image characteristics.
We propose a new data guided generative method for histopathology image segmentation by leveraging the unlabeled data distributions.
Our method is evaluated on glands and nuclei datasets.
arXiv Detail & Related papers (2020-12-17T02:54:19Z) - i-Mix: A Domain-Agnostic Strategy for Contrastive Representation
Learning [117.63815437385321]
We propose i-Mix, a simple yet effective domain-agnostic regularization strategy for improving contrastive representation learning.
In experiments, we demonstrate that i-Mix consistently improves the quality of learned representations across domains.
arXiv Detail & Related papers (2020-10-17T23:32:26Z) - Synthetic Learning: Learn From Distributed Asynchronized Discriminator
GAN Without Sharing Medical Image Data [21.725983290877753]
We propose a data privacy-preserving and communication efficient distributed GAN learning framework named Distributed Asynchronized Discriminator GAN (AsynDGAN)
arXiv Detail & Related papers (2020-05-29T21:05:49Z) - SCAN: Learning to Classify Images without Labels [73.69513783788622]
We advocate a two-step approach where feature learning and clustering are decoupled.
A self-supervised task from representation learning is employed to obtain semantically meaningful features.
We obtain promising results on ImageNet, and outperform several semi-supervised learning methods in the low-data regime.
arXiv Detail & Related papers (2020-05-25T18:12:33Z) - Unshuffling Data for Improved Generalization [65.57124325257409]
Generalization beyond the training distribution is a core challenge in machine learning.
We show that partitioning the data into well-chosen, non-i.i.d. subsets treated as multiple training environments can guide the learning of models with better out-of-distribution generalization.
arXiv Detail & Related papers (2020-02-27T03:07:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.