Semi-Supervised Disentanglement of Class-Related and Class-Independent
Factors in VAE
- URL: http://arxiv.org/abs/2102.00892v1
- Date: Mon, 1 Feb 2021 15:05:24 GMT
- Title: Semi-Supervised Disentanglement of Class-Related and Class-Independent
Factors in VAE
- Authors: Sina Hajimiri, Aryo Lotfi, Mahdieh Soleymani Baghshah
- Abstract summary: We propose a framework capable of disentangling class-related and class-independent factors of variation in data.
Our framework employs an attention mechanism in its latent space in order to improve the process of extracting class-related factors from data.
Experiments show that our framework disentangles class-related and class-independent factors of variation and learns interpretable features.
- Score: 4.533408938245526
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In recent years, extending variational autoencoder's framework to learn
disentangled representations has received much attention. We address this
problem by proposing a framework capable of disentangling class-related and
class-independent factors of variation in data. Our framework employs an
attention mechanism in its latent space in order to improve the process of
extracting class-related factors from data. We also deal with the multimodality
of data distribution by utilizing mixture models as learnable prior
distributions, as well as incorporating the Bhattacharyya coefficient in the
objective function to prevent highly overlapping mixtures. Our model's encoder
is further trained in a semi-supervised manner, with a small fraction of
labeled data, to improve representations' interpretability. Experiments show
that our framework disentangles class-related and class-independent factors of
variation and learns interpretable features. Moreover, we demonstrate our
model's performance with quantitative and qualitative results on various
datasets.
Related papers
- Granularity Matters in Long-Tail Learning [62.30734737735273]
We offer a novel perspective on long-tail learning, inspired by an observation: datasets with finer granularity tend to be less affected by data imbalance.
We introduce open-set auxiliary classes that are visually similar to existing ones, aiming to enhance representation learning for both head and tail classes.
To prevent the overwhelming presence of auxiliary classes from disrupting training, we introduce a neighbor-silencing loss.
arXiv Detail & Related papers (2024-10-21T13:06:21Z) - Influence Functions for Scalable Data Attribution in Diffusion Models [52.92223039302037]
Diffusion models have led to significant advancements in generative modelling.
Yet their widespread adoption poses challenges regarding data attribution and interpretability.
In this paper, we aim to help address such challenges by developing an textitinfluence functions framework.
arXiv Detail & Related papers (2024-10-17T17:59:02Z) - Enhancing Robustness of Foundation Model Representations under
Provenance-related Distribution Shifts [8.298173603769063]
We examine the stability of models based on foundation models under distribution shift.
We focus on confounding by provenance, a form of distribution shift that emerges in the context of multi-institutional datasets.
Results indicate that while foundation models do show some out-of-the-box robustness to confounding-by-provenance related distribution shifts, this can be improved through adjustment.
arXiv Detail & Related papers (2023-12-09T02:02:45Z) - On the Trade-off of Intra-/Inter-class Diversity for Supervised
Pre-training [72.8087629914444]
We study the impact of the trade-off between the intra-class diversity (the number of samples per class) and the inter-class diversity (the number of classes) of a supervised pre-training dataset.
With the size of the pre-training dataset fixed, the best downstream performance comes with a balance on the intra-/inter-class diversity.
arXiv Detail & Related papers (2023-05-20T16:23:50Z) - Leveraging Relational Information for Learning Weakly Disentangled
Representations [11.460692362624533]
Disentanglement is a difficult property to enforce in neural representations.
We present an alternative view over learning (weakly) disentangled representations.
arXiv Detail & Related papers (2022-05-20T09:58:51Z) - Learning Debiased and Disentangled Representations for Semantic
Segmentation [52.35766945827972]
We propose a model-agnostic and training scheme for semantic segmentation.
By randomly eliminating certain class information in each training iteration, we effectively reduce feature dependencies among classes.
Models trained with our approach demonstrate strong results on multiple semantic segmentation benchmarks.
arXiv Detail & Related papers (2021-10-31T16:15:09Z) - Accuracy on the Line: On the Strong Correlation Between
Out-of-Distribution and In-Distribution Generalization [89.73665256847858]
We show that out-of-distribution performance is strongly correlated with in-distribution performance for a wide range of models and distribution shifts.
Specifically, we demonstrate strong correlations between in-distribution and out-of-distribution performance on variants of CIFAR-10 & ImageNet.
We also investigate cases where the correlation is weaker, for instance some synthetic distribution shifts from CIFAR-10-C and the tissue classification dataset Camelyon17-WILDS.
arXiv Detail & Related papers (2021-07-09T19:48:23Z) - Improving the Reconstruction of Disentangled Representation Learners via Multi-Stage Modeling [54.94763543386523]
Current autoencoder-based disentangled representation learning methods achieve disentanglement by penalizing the ( aggregate) posterior to encourage statistical independence of the latent factors.
We present a novel multi-stage modeling approach where the disentangled factors are first learned using a penalty-based disentangled representation learning method.
Then, the low-quality reconstruction is improved with another deep generative model that is trained to model the missing correlated latent variables.
arXiv Detail & Related papers (2020-10-25T18:51:15Z) - Learning Structured Latent Factors from Dependent Data:A Generative
Model Framework from Information-Theoretic Perspective [18.88255368184596]
We present a novel framework for learning generative models with various underlying structures in the latent space.
Our model provides a principled approach to learn a set of semantically meaningful latent factors that reflect various types of desired structures.
arXiv Detail & Related papers (2020-07-21T06:59:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.