Semi-supervised Disentanglement with Independent Vector Variational
Autoencoders
- URL: http://arxiv.org/abs/2003.06581v1
- Date: Sat, 14 Mar 2020 09:20:22 GMT
- Title: Semi-supervised Disentanglement with Independent Vector Variational
Autoencoders
- Authors: Bo-Kyeong Kim, Sungjin Park, Geonmin Kim, Soo-Young Lee
- Abstract summary: We separate generative factors of data into two latent vectors in a variational autoencoder.
To learn the discrete class features, we introduce supervision using a small amount of labeled data.
We show that (i) this vector independence term exists within the result obtained on the evidence decomposing lower bound with multiple latent vectors, and (ii) encouraging such independence along with reducing the total correlation within the vectors enhances disentanglement performance.
- Score: 7.700240949386079
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We aim to separate the generative factors of data into two latent vectors in
a variational autoencoder. One vector captures class factors relevant to target
classification tasks, while the other vector captures style factors relevant to
the remaining information. To learn the discrete class features, we introduce
supervision using a small amount of labeled data, which can simply yet
effectively reduce the effort required for hyperparameter tuning performed in
existing unsupervised methods. Furthermore, we introduce a learning objective
to encourage statistical independence between the vectors. We show that (i)
this vector independence term exists within the result obtained on decomposing
the evidence lower bound with multiple latent vectors, and (ii) encouraging
such independence along with reducing the total correlation within the vectors
enhances disentanglement performance. Experiments conducted on several image
datasets demonstrate that the disentanglement achieved via our method can
improve classification performance and generation controllability.
Related papers
- Disentanglement with Factor Quantized Variational Autoencoders [11.086500036180222]
We propose a discrete variational autoencoder (VAE) based model where the ground truth information about the generative factors are not provided to the model.
We demonstrate the advantages of learning discrete representations over learning continuous representations in facilitating disentanglement.
Our method called FactorQVAE is the first method that combines optimization based disentanglement approaches with discrete representation learning.
arXiv Detail & Related papers (2024-09-23T09:33:53Z) - Spectral Self-supervised Feature Selection [7.052728135831165]
We propose a self-supervised graph-based approach for unsupervised feature selection.
Our method's core involves computing robust pseudo-labels by applying simple processing steps to the graph Laplacian's eigenvectors.
Our approach is shown to be robust to challenging scenarios, such as the presence of outliers and complex substructures.
arXiv Detail & Related papers (2024-07-12T07:29:08Z) - Knowledge Composition using Task Vectors with Learned Anisotropic Scaling [51.4661186662329]
We introduce aTLAS, an algorithm that linearly combines parameter blocks with different learned coefficients, resulting in anisotropic scaling at the task vector level.
We show that such linear combinations explicitly exploit the low intrinsicity of pre-trained models, with only a few coefficients being the learnable parameters.
We demonstrate the effectiveness of our method in task arithmetic, few-shot recognition and test-time adaptation, with supervised or unsupervised objectives.
arXiv Detail & Related papers (2024-07-03T07:54:08Z) - Convolutional autoencoder-based multimodal one-class classification [80.52334952912808]
One-class classification refers to approaches of learning using data from a single class only.
We propose a deep learning one-class classification method suitable for multimodal data.
arXiv Detail & Related papers (2023-09-25T12:31:18Z) - Adaptive Discrete Communication Bottlenecks with Dynamic Vector
Quantization [76.68866368409216]
We propose learning to dynamically select discretization tightness conditioned on inputs.
We show that dynamically varying tightness in communication bottlenecks can improve model performance on visual reasoning and reinforcement learning tasks.
arXiv Detail & Related papers (2022-02-02T23:54:26Z) - Is Disentanglement enough? On Latent Representations for Controllable
Music Generation [78.8942067357231]
In the absence of a strong generative decoder, disentanglement does not necessarily imply controllability.
The structure of the latent space with respect to the VAE-decoder plays an important role in boosting the ability of a generative model to manipulate different attributes.
arXiv Detail & Related papers (2021-08-01T18:37:43Z) - Semi-Supervised Disentanglement of Class-Related and Class-Independent
Factors in VAE [4.533408938245526]
We propose a framework capable of disentangling class-related and class-independent factors of variation in data.
Our framework employs an attention mechanism in its latent space in order to improve the process of extracting class-related factors from data.
Experiments show that our framework disentangles class-related and class-independent factors of variation and learns interpretable features.
arXiv Detail & Related papers (2021-02-01T15:05:24Z) - On Supervised Classification of Feature Vectors with Independent and
Non-Identically Distributed Elements [10.52087851034255]
We investigate the problem of classifying feature vectors with mutually independent but non-identically distributed elements.
We show that the error probability goes to zero as the length of the feature vectors grows, even when there is only one training feature vector per label available.
arXiv Detail & Related papers (2020-08-01T06:49:50Z) - Unsupervised Controllable Generation with Self-Training [90.04287577605723]
controllable generation with GANs remains a challenging research problem.
We propose an unsupervised framework to learn a distribution of latent codes that control the generator through self-training.
Our framework exhibits better disentanglement compared to other variants such as the variational autoencoder.
arXiv Detail & Related papers (2020-07-17T21:50:35Z) - Self-supervised Video Object Segmentation [76.83567326586162]
The objective of this paper is self-supervised representation learning, with the goal of solving semi-supervised video object segmentation (a.k.a. dense tracking)
We make the following contributions: (i) we propose to improve the existing self-supervised approach, with a simple, yet more effective memory mechanism for long-term correspondence matching; (ii) by augmenting the self-supervised approach with an online adaptation module, our method successfully alleviates tracker drifts caused by spatial-temporal discontinuity; (iv) we demonstrate state-of-the-art results among the self-supervised approaches on DAVIS-2017 and YouTube
arXiv Detail & Related papers (2020-06-22T17:55:59Z) - Generalized Adversarially Learned Inference [42.40405470084505]
We develop methods of inference of latent variables in GANs by adversarially training an image generator along with an encoder to match two joint distributions of image and latent vector pairs.
We incorporate multiple layers of feedback on reconstructions, self-supervision, and other forms of supervision based on prior or learned knowledge about the desired solutions.
arXiv Detail & Related papers (2020-06-15T02:18:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.