CUDA-GR: Controllable Unsupervised Domain Adaptation for Gaze
Redirection
- URL: http://arxiv.org/abs/2106.10852v1
- Date: Mon, 21 Jun 2021 04:39:42 GMT
- Title: CUDA-GR: Controllable Unsupervised Domain Adaptation for Gaze
Redirection
- Authors: Swati Jindal, Xin Eric Wang
- Abstract summary: The aim of gaze redirection is to manipulate the gaze in an image to the desired direction.
Advancement in generative adversarial networks has shown excellent results in generating photo-realistic images.
To enable such fine-tuned control, one needs to obtain ground truth annotations for the training data which can be very expensive.
- Score: 3.0141238193080295
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The aim of gaze redirection is to manipulate the gaze in an image to the
desired direction. However, existing methods are inadequate in generating
perceptually reasonable images. Advancement in generative adversarial networks
has shown excellent results in generating photo-realistic images. Though, they
still lack the ability to provide finer control over different image
attributes. To enable such fine-tuned control, one needs to obtain ground truth
annotations for the training data which can be very expensive. In this paper,
we propose an unsupervised domain adaptation framework, called CUDA-GR, that
learns to disentangle gaze representations from the labeled source domain and
transfers them to an unlabeled target domain. Our method enables fine-grained
control over gaze directions while preserving the appearance information of the
person. We show that the generated image-labels pairs in the target domain are
effective in knowledge transfer and can boost the performance of the downstream
tasks. Extensive experiments on the benchmarking datasets show that the
proposed method can outperform state-of-the-art techniques in both quantitative
and qualitative evaluation.
Related papers
- SiamSeg: Self-Training with Contrastive Learning for Unsupervised Domain Adaptation Semantic Segmentation in Remote Sensing [14.007392647145448]
UDA enables models to learn from unlabeled target domain data while training on labeled source domain data.
We propose integrating contrastive learning into UDA, enhancing the model's capacity to capture semantic information.
Our SimSeg method outperforms existing approaches, achieving state-of-the-art results.
arXiv Detail & Related papers (2024-10-17T11:59:39Z) - TraDiffusion: Trajectory-Based Training-Free Image Generation [85.39724878576584]
We propose a training-free, trajectory-based controllable T2I approach, termed TraDiffusion.
This novel method allows users to effortlessly guide image generation via mouse trajectories.
arXiv Detail & Related papers (2024-08-19T07:01:43Z) - Learning Gaze-aware Compositional GAN [30.714854907472333]
We present a generative framework to create annotated gaze data by leveraging the benefits of labeled and unlabeled data sources.
We show our approach's effectiveness in generating within-domain image augmentations in the ETH-XGaze dataset and cross-domain augmentations in the CelebAMask-HQ dataset domain for gaze estimation training.
arXiv Detail & Related papers (2024-05-31T07:07:54Z) - Tell, Don't Show!: Language Guidance Eases Transfer Across Domains in Images and Videos [69.29778009769862]
We introduce LaGTran, a framework that guides robust transfer of discriminative knowledge from labeled source to unlabeled target data with domain gaps.
Motivated by our observation that semantically richer text modality has more favorable transfer properties, we devise a transfer mechanism to use a source-trained text-classifier to generate predictions on the target text descriptions.
Our approach driven by language guidance is surprisingly easy and simple, yet significantly outperforms all prior approaches on challenging datasets like GeoNet and DomainNet.
arXiv Detail & Related papers (2024-03-08T18:58:46Z) - Unsupervised Discovery of Disentangled Manifolds in GANs [74.24771216154105]
Interpretable generation process is beneficial to various image editing applications.
We propose a framework to discover interpretable directions in the latent space given arbitrary pre-trained generative adversarial networks.
arXiv Detail & Related papers (2020-11-24T02:18:08Z) - Self-Learning Transformations for Improving Gaze and Head Redirection [49.61091281780071]
We propose a novel generative model for images of faces, that is capable of producing high-quality images under fine-grained control over eye gaze and head orientation angles.
This requires the disentangling of many appearance related factors including gaze and head orientation but also lighting, hue etc.
We show that explicitly disentangling task-irrelevant factors results in more accurate modelling of gaze and head orientation.
arXiv Detail & Related papers (2020-10-23T11:18:37Z) - Controllable Continuous Gaze Redirection [47.15883248953411]
We present interpGaze, a novel framework for controllable gaze redirection.
Our goal is to redirect the eye gaze of one person into any gaze direction depicted in the reference image.
The proposed interpGaze outperforms state-of-the-art methods in terms of image quality and redirection precision.
arXiv Detail & Related papers (2020-10-09T11:50:06Z) - Coarse-to-Fine Gaze Redirection with Numerical and Pictorial Guidance [74.27389895574422]
We propose a novel gaze redirection framework which exploits both a numerical and a pictorial direction guidance.
The proposed method outperforms the state-of-the-art approaches in terms of both image quality and redirection precision.
arXiv Detail & Related papers (2020-04-07T01:17:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.