Do Generative Models Know Disentanglement? Contrastive Learning is All
You Need
- URL: http://arxiv.org/abs/2102.10543v1
- Date: Sun, 21 Feb 2021 08:01:20 GMT
- Title: Do Generative Models Know Disentanglement? Contrastive Learning is All
You Need
- Authors: Xuanchi Ren, Tao Yang, Yuwang Wang, Wenjun Zeng
- Abstract summary: We propose an unsupervised and model-agnostic method: Disentanglement via Contrast (DisCo) in the Variation Space.
DisCo achieves the state-of-the-art disentanglement given pretrained non-disentangled generative models, including GAN, VAE, and Flow.
- Score: 59.033559925639075
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Disentangled generative models are typically trained with an extra
regularization term, which encourages the traversal of each latent factor to
make a distinct and independent change at the cost of generation quality. When
traversing the latent space of generative models trained without the
disentanglement term, the generated samples show semantically meaningful
change, raising the question: do generative models know disentanglement? We
propose an unsupervised and model-agnostic method: Disentanglement via Contrast
(DisCo) in the Variation Space. DisCo consists of: (i) a Navigator providing
traversal directions in the latent space, and (ii) a $\Delta$-Contrastor
composed of two shared-weight Encoders, which encode image pairs along these
directions to disentangled representations respectively, and a difference
operator to map the encoded representations to the Variation Space. We propose
two more key techniques for DisCo: entropy-based domination loss to make the
encoded representations more disentangled and the strategy of flipping hard
negatives to address directions with the same semantic meaning. By optimizing
the Navigator to discover disentangled directions in the latent space and
Encoders to extract disentangled representations from images with Contrastive
Learning, DisCo achieves the state-of-the-art disentanglement given pretrained
non-disentangled generative models, including GAN, VAE, and Flow. Project page
at https://github.com/xrenaa/DisCo.
Related papers
- Complexity Matters: Rethinking the Latent Space for Generative Modeling [65.64763873078114]
In generative modeling, numerous successful approaches leverage a low-dimensional latent space, e.g., Stable Diffusion.
In this study, we aim to shed light on this under-explored topic by rethinking the latent space from the perspective of model complexity.
arXiv Detail & Related papers (2023-07-17T07:12:29Z) - Spatial Steerability of GANs via Self-Supervision from Discriminator [123.27117057804732]
We propose a self-supervised approach to improve the spatial steerability of GANs without searching for steerable directions in the latent space.
Specifically, we design randomly sampled Gaussian heatmaps to be encoded into the intermediate layers of generative models as spatial inductive bias.
During inference, users can interact with the spatial heatmaps in an intuitive manner, enabling them to edit the output image by adjusting the scene layout, moving, or removing objects.
arXiv Detail & Related papers (2023-01-20T07:36:29Z) - Decoupled Multi-task Learning with Cyclical Self-Regulation for Face
Parsing [71.19528222206088]
We propose a novel Decoupled Multi-task Learning with Cyclical Self-Regulation for face parsing.
Specifically, DML-CSR designs a multi-task model which comprises face parsing, binary edge, and category edge detection.
Our method achieves the new state-of-the-art performance on the Helen, CelebA-HQ, and LapaMask datasets.
arXiv Detail & Related papers (2022-03-28T02:12:30Z) - The Transitive Information Theory and its Application to Deep Generative
Models [0.0]
Variational Autoencoder (VAE) could be pushed in two opposite directions.
Existing methods narrow the issues to the rate-distortion trade-off between compression and reconstruction.
We develop a system that learns a hierarchy of disentangled representation together with a mechanism for recombining the learned representation for generalization.
arXiv Detail & Related papers (2022-03-09T22:35:02Z) - Orthogonal Jacobian Regularization for Unsupervised Disentanglement in
Image Generation [64.92152574895111]
We propose a simple Orthogonal Jacobian Regularization (OroJaR) to encourage deep generative model to learn disentangled representations.
Our method is effective in disentangled and controllable image generation, and performs favorably against the state-of-the-art methods.
arXiv Detail & Related papers (2021-08-17T15:01:46Z) - Mask-Guided Discovery of Semantic Manifolds in Generative Models [0.0]
StyleGAN2 generates images of human faces from random vectors in a lower-dimensional latent space.
The model behaves as a black box, providing neither control over its output nor insight into the structures it has learned from the data.
We present a method to explore the manifold of changes of spatially localized regions of the face.
arXiv Detail & Related papers (2021-05-15T18:06:38Z) - Unsupervised Controllable Generation with Self-Training [90.04287577605723]
controllable generation with GANs remains a challenging research problem.
We propose an unsupervised framework to learn a distribution of latent codes that control the generator through self-training.
Our framework exhibits better disentanglement compared to other variants such as the variational autoencoder.
arXiv Detail & Related papers (2020-07-17T21:50:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.