CoordGAN: Self-Supervised Dense Correspondences Emerge from GANs
- URL: http://arxiv.org/abs/2203.16521v1
- Date: Wed, 30 Mar 2022 17:55:09 GMT
- Title: CoordGAN: Self-Supervised Dense Correspondences Emerge from GANs
- Authors: Jiteng Mu, Shalini De Mello, Zhiding Yu, Nuno Vasconcelos, Xiaolong
Wang, Jan Kautz, Sifei Liu
- Abstract summary: We introduce Coordinate GAN (CoordGAN), a structure-texture disentangled GAN that learns a dense correspondence map for each generated image.
We show that the proposed generator achieves better structure and texture disentanglement compared to existing approaches.
- Score: 129.51129173514502
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent advances show that Generative Adversarial Networks (GANs) can
synthesize images with smooth variations along semantically meaningful latent
directions, such as pose, expression, layout, etc. While this indicates that
GANs implicitly learn pixel-level correspondences across images, few studies
explored how to extract them explicitly. In this work, we introduce Coordinate
GAN (CoordGAN), a structure-texture disentangled GAN that learns a dense
correspondence map for each generated image. We represent the correspondence
maps of different images as warped coordinate frames transformed from a
canonical coordinate frame, i.e., the correspondence map, which describes the
structure (e.g., the shape of a face), is controlled via a transformation.
Hence, finding correspondences boils down to locating the same coordinate in
different correspondence maps. In CoordGAN, we sample a transformation to
represent the structure of a synthesized instance, while an independent texture
branch is responsible for rendering appearance details orthogonal to the
structure. Our approach can also extract dense correspondence maps for real
images by adding an encoder on top of the generator. We quantitatively
demonstrate the quality of the learned dense correspondences through
segmentation mask transfer on multiple datasets. We also show that the proposed
generator achieves better structure and texture disentanglement compared to
existing approaches. Project page: https://jitengmu.github.io/CoordGAN/
Related papers
- Transforming Visual Scene Graphs to Image Captions [69.13204024990672]
We propose to transform Scene Graphs (TSG) into more descriptive captions.
In TSG, we apply multi-head attention (MHA) to design the Graph Neural Network (GNN) for embedding scene graphs.
In TSG, each expert is built on MHA, for discriminating the graph embeddings to generate different kinds of words.
arXiv Detail & Related papers (2023-05-03T15:18:37Z) - Extracting Semantic Knowledge from GANs with Unsupervised Learning [65.32631025780631]
Generative Adversarial Networks (GANs) encode semantics in feature maps in a linearly separable form.
We propose a novel clustering algorithm, named KLiSH, which leverages the linear separability to cluster GAN's features.
KLiSH succeeds in extracting fine-grained semantics of GANs trained on datasets of various objects.
arXiv Detail & Related papers (2022-11-30T03:18:16Z) - Diffusion-Based Scene Graph to Image Generation with Masked Contrastive
Pre-Training [112.94542676251133]
We propose to learn scene graph embeddings by directly optimizing their alignment with images.
Specifically, we pre-train an encoder to extract both global and local information from scene graphs.
The resulting method, called SGDiff, allows for the semantic manipulation of generated images by modifying scene graph nodes and connections.
arXiv Detail & Related papers (2022-11-21T01:11:19Z) - Generative Subgraph Contrast for Self-Supervised Graph Representation
Learning [16.374143635724327]
We propose a novel adaptive subgraph generation based contrastive learning framework for efficient and robust self-supervised graph representation learning.
It aims to generate contrastive samples by capturing the intrinsic structures of the graph and distinguish the samples based on the features and structures of subgraphs simultaneously.
arXiv Detail & Related papers (2022-07-25T09:08:46Z) - Learning Implicit Feature Alignment Function for Semantic Segmentation [51.36809814890326]
Implicit Feature Alignment function (IFA) is inspired by the rapidly expanding topic of implicit neural representations.
We show that IFA implicitly aligns the feature maps at different levels and is capable of producing segmentation maps in arbitrary resolutions.
Our method can be combined with improvement on various architectures, and it achieves state-of-the-art accuracy trade-off on common benchmarks.
arXiv Detail & Related papers (2022-06-17T09:40:14Z) - Linear Semantics in Generative Adversarial Networks [26.123252503846942]
We aim to better understand the semantic representation of GANs, and enable semantic control in GAN's generation process.
We find that a well-trained GAN encodes image semantics in its internal feature maps in a surprisingly simple way.
We propose two few-shot image editing approaches, namely Semantic-Conditional Sampling and Semantic Image Editing.
arXiv Detail & Related papers (2021-04-01T14:18:48Z) - The Geometry of Deep Generative Image Models and its Applications [0.0]
Generative adversarial networks (GANs) have emerged as a powerful unsupervised method to model the statistical patterns of real-world data sets.
These networks are trained to map random inputs in their latent space to new samples representative of the learned data.
The structure of the latent space is hard to intuit due to its high dimensionality and the non-linearity of the generator.
arXiv Detail & Related papers (2021-01-15T07:57:33Z) - Structural-analogy from a Single Image Pair [118.61885732829117]
In this paper, we explore the capabilities of neural networks to understand image structure given only a single pair of images, A and B.
We generate an image that keeps the appearance and style of B, but has a structural arrangement that corresponds to A.
Our method can be used to generate high quality imagery in other conditional generation tasks utilizing images A and B only.
arXiv Detail & Related papers (2020-04-05T14:51:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.