Face Identity Disentanglement via Latent Space Mapping
- URL: http://arxiv.org/abs/2005.07728v3
- Date: Mon, 19 Oct 2020 12:24:42 GMT
- Title: Face Identity Disentanglement via Latent Space Mapping
- Authors: Yotam Nitzan, Amit Bermano, Yangyan Li, Daniel Cohen-Or
- Abstract summary: We present a method that learns how to represent data in a disentangled way, with minimal supervision.
Our key insight is to decouple the processes of disentanglement and synthesis, by employing a leading pre-trained unconditional image generator, such as StyleGAN.
We show that our method successfully disentangles identity from other facial attributes, surpassing existing methods.
- Score: 47.27253184341152
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning disentangled representations of data is a fundamental problem in
artificial intelligence. Specifically, disentangled latent representations
allow generative models to control and compose the disentangled factors in the
synthesis process. Current methods, however, require extensive supervision and
training, or instead, noticeably compromise quality. In this paper, we present
a method that learns how to represent data in a disentangled way, with minimal
supervision, manifested solely using available pre-trained networks. Our key
insight is to decouple the processes of disentanglement and synthesis, by
employing a leading pre-trained unconditional image generator, such as
StyleGAN. By learning to map into its latent space, we leverage both its
state-of-the-art quality, and its rich and expressive latent space, without the
burden of training it. We demonstrate our approach on the complex and high
dimensional domain of human heads. We evaluate our method qualitatively and
quantitatively, and exhibit its success with de-identification operations and
with temporal identity coherency in image sequences. Through extensive
experimentation, we show that our method successfully disentangles identity
from other facial attributes, surpassing existing methods, even though they
require more training and supervision.
Related papers
- Self-Explainable Affordance Learning with Embodied Caption [63.88435741872204]
We introduce Self-Explainable Affordance learning (SEA) with embodied caption.
SEA enables robots to articulate their intentions and bridge the gap between explainable vision-language caption and visual affordance learning.
We propose a novel model to effectively combine affordance grounding with self-explanation in a simple but efficient manner.
arXiv Detail & Related papers (2024-04-08T15:22:38Z) - Multi Positive Contrastive Learning with Pose-Consistent Generated Images [0.873811641236639]
We propose the generation of visually distinct images with identical human poses.
We then propose a novel multi-positive contrastive learning, which optimally utilize the previously generated images.
Despite using only less than 1% amount of data compared to current state-of-the-art method, GenPoCCL captures structural features of the human body more effectively.
arXiv Detail & Related papers (2024-04-04T07:26:26Z) - Segue: Side-information Guided Generative Unlearnable Examples for
Facial Privacy Protection in Real World [64.4289385463226]
We propose Segue: Side-information guided generative unlearnable examples.
To improve transferability, we introduce side information such as true labels and pseudo labels.
It can resist JPEG compression, adversarial training, and some standard data augmentations.
arXiv Detail & Related papers (2023-10-24T06:22:37Z) - URLOST: Unsupervised Representation Learning without Stationarity or
Topology [26.17135629579595]
We introduce a novel framework that learns from high-dimensional data lacking stationarity and topology.
Our model combines a learnable self-organizing layer, density adjusted spectral clustering, and masked autoencoders.
We evaluate its effectiveness on simulated biological vision data, neural recordings from the primary visual cortex, and gene expression datasets.
arXiv Detail & Related papers (2023-10-06T18:00:02Z) - Free-ATM: Exploring Unsupervised Learning on Diffusion-Generated Images
with Free Attention Masks [64.67735676127208]
Text-to-image diffusion models have shown great potential for benefiting image recognition.
Although promising, there has been inadequate exploration dedicated to unsupervised learning on diffusion-generated images.
We introduce customized solutions by fully exploiting the aforementioned free attention masks.
arXiv Detail & Related papers (2023-08-13T10:07:46Z) - Attribute-preserving Face Dataset Anonymization via Latent Code
Optimization [64.4569739006591]
We present a task-agnostic anonymization procedure that directly optimize the images' latent representation in the latent space of a pre-trained GAN.
We demonstrate through a series of experiments that our method is capable of anonymizing the identity of the images whilst -- crucially -- better-preserving the facial attributes.
arXiv Detail & Related papers (2023-03-20T17:34:05Z) - Self-supervised Transformer for Deepfake Detection [112.81127845409002]
Deepfake techniques in real-world scenarios require stronger generalization abilities of face forgery detectors.
Inspired by transfer learning, neural networks pre-trained on other large-scale face-related tasks may provide useful features for deepfake detection.
In this paper, we propose a self-supervised transformer based audio-visual contrastive learning method.
arXiv Detail & Related papers (2022-03-02T17:44:40Z) - Learning a Deep Reinforcement Learning Policy Over the Latent Space of a
Pre-trained GAN for Semantic Age Manipulation [4.306143768014157]
We learn a conditional policy for semantic manipulation along specific attributes under defined identity bounds.
Results show that our learned policy samples high fidelity images with required age alterations.
arXiv Detail & Related papers (2020-11-02T13:15:18Z) - Disentangled and Controllable Face Image Generation via 3D
Imitative-Contrastive Learning [43.53235319568048]
We propose DiscoFaceGAN, an approach for face image generation of virtual people with disentangled, precisely-controllable latent representations.
We embed 3D priors into adversarial learning and train the network to imitate the image formation of an analytic 3D face deformation and rendering process.
arXiv Detail & Related papers (2020-04-24T11:21:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.