Representation Decomposition for Image Manipulation and Beyond
- URL: http://arxiv.org/abs/2011.00788v2
- Date: Wed, 23 Mar 2022 09:23:59 GMT
- Title: Representation Decomposition for Image Manipulation and Beyond
- Authors: Shang-Fu Chen, Jia-Wei Yan, Ya-Fan Su, Yu-Chiang Frank Wang
- Abstract summary: decomposition-GAN (dec-GAN) is able to achieve the decomposition of an existing latent representation into content and attribute features.
Our experiments on multiple image datasets confirm the effectiveness and robustness of our dec-GAN over recent representation disentanglement models.
- Score: 29.991777603295816
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Representation disentanglement aims at learning interpretable features, so
that the output can be recovered or manipulated accordingly. While existing
works like infoGAN and AC-GAN exist, they choose to derive disjoint attribute
code for feature disentanglement, which is not applicable for existing/trained
generative models. In this paper, we propose a decomposition-GAN (dec-GAN),
which is able to achieve the decomposition of an existing latent representation
into content and attribute features. Guided by the classifier pre-trained on
the attributes of interest, our dec-GAN decomposes the attributes of interest
from the latent representation, while data recovery and feature consistency
objectives enforce the learning of our proposed method. Our experiments on
multiple image datasets confirm the effectiveness and robustness of our dec-GAN
over recent representation disentanglement models.
Related papers
- An Information Compensation Framework for Zero-Shot Skeleton-based Action Recognition [49.45660055499103]
Zero-shot human skeleton-based action recognition aims to construct a model that can recognize actions outside the categories seen during training.
Previous research has focused on aligning sequences' visual and semantic spatial distributions.
We introduce a new loss function sampling method to obtain a tight and robust representation.
arXiv Detail & Related papers (2024-06-02T06:53:01Z) - Weakly-supervised Contrastive Learning for Unsupervised Object Discovery [52.696041556640516]
Unsupervised object discovery is promising due to its ability to discover objects in a generic manner.
We design a semantic-guided self-supervised learning model to extract high-level semantic features from images.
We introduce Principal Component Analysis (PCA) to localize object regions.
arXiv Detail & Related papers (2023-07-07T04:03:48Z) - Learning disentangled representations for explainable chest X-ray
classification using Dirichlet VAEs [68.73427163074015]
This study explores the use of the Dirichlet Variational Autoencoder (DirVAE) for learning disentangled latent representations of chest X-ray (CXR) images.
The predictive capacity of multi-modal latent representations learned by DirVAE models is investigated through implementation of an auxiliary multi-label classification task.
arXiv Detail & Related papers (2023-02-06T18:10:08Z) - Multi-Attribute Balanced Sampling for Disentangled GAN Controls [0.0]
Various controls over the generated data can be extracted from the latent space of a pre-trained GAN.
We show that this approach outperforms state-of-the-art classifier-based methods while avoiding the need for disentanglement-enforcing post-processing.
arXiv Detail & Related papers (2021-10-28T08:44:13Z) - GLOWin: A Flow-based Invertible Generative Framework for Learning
Disentangled Feature Representations in Medical Images [40.58581577183134]
Flow-based generative models have been proposed to generate realistic images by directly modeling the data distribution with invertible functions.
We propose a new flow-based generative model framework, named GLOWin, that is end-to-end invertible and able to learn disentangled representations.
arXiv Detail & Related papers (2021-03-19T15:47:01Z) - Out-of-distribution Generalization via Partial Feature Decorrelation [72.96261704851683]
We present a novel Partial Feature Decorrelation Learning (PFDL) algorithm, which jointly optimize a feature decomposition network and the target image classification model.
The experiments on real-world datasets demonstrate that our method can improve the backbone model's accuracy on OOD image classification datasets.
arXiv Detail & Related papers (2020-07-30T05:48:48Z) - Generative Hierarchical Features from Synthesizing Images [65.66756821069124]
We show that learning to synthesize images can bring remarkable hierarchical visual features that are generalizable across a wide range of applications.
The visual feature produced by our encoder, termed as Generative Hierarchical Feature (GH-Feat), has strong transferability to both generative and discriminative tasks.
arXiv Detail & Related papers (2020-07-20T18:04:14Z) - Two-Level Adversarial Visual-Semantic Coupling for Generalized Zero-shot
Learning [21.89909688056478]
We propose a new two-level joint idea to augment the generative network with an inference network during training.
This provides strong cross-modal interaction for effective transfer of knowledge between visual and semantic domains.
We evaluate our approach on four benchmark datasets against several state-of-the-art methods, and show its performance.
arXiv Detail & Related papers (2020-07-15T15:34:09Z) - Closed-Form Factorization of Latent Semantics in GANs [65.42778970898534]
A rich set of interpretable dimensions has been shown to emerge in the latent space of the Generative Adversarial Networks (GANs) trained for synthesizing images.
In this work, we examine the internal representation learned by GANs to reveal the underlying variation factors in an unsupervised manner.
We propose a closed-form factorization algorithm for latent semantic discovery by directly decomposing the pre-trained weights.
arXiv Detail & Related papers (2020-07-13T18:05:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.