Orthogonal Jacobian Regularization for Unsupervised Disentanglement in
Image Generation
- URL: http://arxiv.org/abs/2108.07668v1
- Date: Tue, 17 Aug 2021 15:01:46 GMT
- Title: Orthogonal Jacobian Regularization for Unsupervised Disentanglement in
Image Generation
- Authors: Yuxiang Wei, Yupeng Shi, Xiao Liu, Zhilong Ji, Yuan Gao, Zhongqin Wu,
Wangmeng Zuo
- Abstract summary: We propose a simple Orthogonal Jacobian Regularization (OroJaR) to encourage deep generative model to learn disentangled representations.
Our method is effective in disentangled and controllable image generation, and performs favorably against the state-of-the-art methods.
- Score: 64.92152574895111
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unsupervised disentanglement learning is a crucial issue for understanding
and exploiting deep generative models. Recently, SeFa tries to find latent
disentangled directions by performing SVD on the first projection of a
pre-trained GAN. However, it is only applied to the first layer and works in a
post-processing way. Hessian Penalty minimizes the off-diagonal entries of the
output's Hessian matrix to facilitate disentanglement, and can be applied to
multi-layers.However, it constrains each entry of output independently, making
it not sufficient in disentangling the latent directions (e.g., shape, size,
rotation, etc.) of spatially correlated variations. In this paper, we propose a
simple Orthogonal Jacobian Regularization (OroJaR) to encourage deep generative
model to learn disentangled representations. It simply encourages the variation
of output caused by perturbations on different latent dimensions to be
orthogonal, and the Jacobian with respect to the input is calculated to
represent this variation. We show that our OroJaR also encourages the output's
Hessian matrix to be diagonal in an indirect manner. In contrast to the Hessian
Penalty, our OroJaR constrains the output in a holistic way, making it very
effective in disentangling latent dimensions corresponding to spatially
correlated variations. Quantitative and qualitative experimental results show
that our method is effective in disentangled and controllable image generation,
and performs favorably against the state-of-the-art methods. Our code is
available at https://github.com/csyxwei/OroJaR
Related papers
- Tilt your Head: Activating the Hidden Spatial-Invariance of Classifiers [0.7704032792820767]
Deep neural networks are applied in more and more areas of everyday life.
They still lack essential abilities, such as robustly dealing with spatially transformed input signals.
We propose a novel technique to emulate such an inference process for neural nets.
arXiv Detail & Related papers (2024-05-06T09:47:29Z) - Unsupervised Discovery of Interpretable Directions in h-space of
Pre-trained Diffusion Models [63.1637853118899]
We propose the first unsupervised and learning-based method to identify interpretable directions in h-space of pre-trained diffusion models.
We employ a shift control module that works on h-space of pre-trained diffusion models to manipulate a sample into a shifted version of itself.
By jointly optimizing them, the model will spontaneously discover disentangled and interpretable directions.
arXiv Detail & Related papers (2023-10-15T18:44:30Z) - Improving Diffusion Models for Inverse Problems using Manifold Constraints [55.91148172752894]
We show that current solvers throw the sample path off the data manifold, and hence the error accumulates.
To address this, we propose an additional correction term inspired by the manifold constraint.
We show that our method is superior to the previous methods both theoretically and empirically.
arXiv Detail & Related papers (2022-06-02T09:06:10Z) - Unsupervised Disentanglement with Tensor Product Representations on the
Torus [78.6315881294899]
Current methods for learning representations with auto-encoders almost exclusively employ vectors as the latent representations.
In this work, we propose to employ a tensor product structure for this purpose.
In contrast to the conventional variations methods, which are targeted toward normally distributed features, the latent space in our representation is distributed uniformly over a set of unit circles.
arXiv Detail & Related papers (2022-02-13T04:23:12Z) - LARGE: Latent-Based Regression through GAN Semantics [42.50535188836529]
We propose a novel method for solving regression tasks using few-shot or weak supervision.
We show that our method can be applied across a wide range of domains, leverage multiple latent direction discovery frameworks, and achieve state-of-the-art results.
arXiv Detail & Related papers (2021-07-22T17:55:35Z) - Do Generative Models Know Disentanglement? Contrastive Learning is All
You Need [59.033559925639075]
We propose an unsupervised and model-agnostic method: Disentanglement via Contrast (DisCo) in the Variation Space.
DisCo achieves the state-of-the-art disentanglement given pretrained non-disentangled generative models, including GAN, VAE, and Flow.
arXiv Detail & Related papers (2021-02-21T08:01:20Z) - Unsupervised Controllable Generation with Self-Training [90.04287577605723]
controllable generation with GANs remains a challenging research problem.
We propose an unsupervised framework to learn a distribution of latent codes that control the generator through self-training.
Our framework exhibits better disentanglement compared to other variants such as the variational autoencoder.
arXiv Detail & Related papers (2020-07-17T21:50:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.