DeepPortraitDrawing: Generating Human Body Images from Freehand Sketches
- URL: http://arxiv.org/abs/2205.02070v1
- Date: Wed, 4 May 2022 14:02:45 GMT
- Title: DeepPortraitDrawing: Generating Human Body Images from Freehand Sketches
- Authors: Xian Wu, Chen Wang, Hongbo Fu, Ariel Shamir, Song-Hai Zhang, Shi-Min
Hu
- Abstract summary: We present DeepDrawing, a framework for converting roughly drawn sketches to realistic human body images.
To encode complicated body shapes under various poses, we take a local-to-global approach.
Our method produces more realistic images than the state-of-the-art sketch-to-image synthesis techniques.
- Score: 75.4318318890065
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Researchers have explored various ways to generate realistic images from
freehand sketches, e.g., for objects and human faces. However, how to generate
realistic human body images from sketches is still a challenging problem. It
is, first because of the sensitivity to human shapes, second because of the
complexity of human images caused by body shape and pose changes, and third
because of the domain gap between realistic images and freehand sketches. In
this work, we present DeepPortraitDrawing, a deep generative framework for
converting roughly drawn sketches to realistic human body images. To encode
complicated body shapes under various poses, we take a local-to-global
approach. Locally, we employ semantic part auto-encoders to construct
part-level shape spaces, which are useful for refining the geometry of an input
pre-segmented hand-drawn sketch. Globally, we employ a cascaded spatial
transformer network to refine the structure of body parts by adjusting their
spatial locations and relative proportions. Finally, we use a global synthesis
network for the sketch-to-image translation task, and a face refinement network
to enhance facial details. Extensive experiments have shown that given roughly
sketched human portraits, our method produces more realistic images than the
state-of-the-art sketch-to-image synthesis techniques.
Related papers
- Sketch2Human: Deep Human Generation with Disentangled Geometry and Appearance Control [27.23770287587972]
This work presents Sketch2Human, the first system for controllable full-body human image generation guided by a semantic sketch.
We present a sketch encoder trained with a large synthetic dataset sampled from StyleGAN-Human's latent space.
Although our method is trained with synthetic data, it can handle hand-drawn sketches as well.
arXiv Detail & Related papers (2024-04-24T14:24:57Z) - Human-Inspired Facial Sketch Synthesis with Dynamic Adaptation [25.293899668984018]
Facial sketch synthesis (FSS) aims to generate a vivid sketch portrait from a given facial photo.
In this paper, we propose a novel Human-Inspired Dynamic Adaptation (HIDA) method.
We show that HIDA can generate high-quality sketches in multiple styles, and significantly outperforms previous methods.
arXiv Detail & Related papers (2023-09-01T02:27:05Z) - Picture that Sketch: Photorealistic Image Generation from Abstract
Sketches [109.69076457732632]
Given an abstract, deformed, ordinary sketch from untrained amateurs like you and me, this paper turns it into a photorealistic image.
We do not dictate an edgemap-like sketch to start with, but aim to work with abstract free-hand human sketches.
In doing so, we essentially democratise the sketch-to-photo pipeline, "picturing" a sketch regardless of how good you sketch.
arXiv Detail & Related papers (2023-03-20T14:49:03Z) - Pose Guided Human Image Synthesis with Partially Decoupled GAN [25.800174118151638]
Pose Guided Human Image Synthesis (PGHIS) is a challenging task of transforming a human image from the reference pose to a target pose.
We propose a method by decoupling the human body into several parts to guide the synthesis of a realistic image of the person.
In addition, we design a multi-head attention-based module for PGHIS.
arXiv Detail & Related papers (2022-10-07T15:31:37Z) - Animatable Neural Radiance Fields from Monocular RGB Video [72.6101766407013]
We present animatable neural radiance fields for detailed human avatar creation from monocular videos.
Our approach extends neural radiance fields to the dynamic scenes with human movements via introducing explicit pose-guided deformation.
In experiments we show that the proposed approach achieves 1) implicit human geometry and appearance reconstruction with high-quality details, 2) photo-realistic rendering of the human from arbitrary views, and 3) animation of the human with arbitrary poses.
arXiv Detail & Related papers (2021-06-25T13:32:23Z) - PISE: Person Image Synthesis and Editing with Decoupled GAN [64.70360318367943]
We propose PISE, a novel two-stage generative model for Person Image Synthesis and Editing.
For human pose transfer, we first synthesize a human parsing map aligned with the target pose to represent the shape of clothing.
To decouple the shape and style of clothing, we propose joint global and local per-region encoding and normalization.
arXiv Detail & Related papers (2021-03-06T04:32:06Z) - Style and Pose Control for Image Synthesis of Humans from a Single
Monocular View [78.6284090004218]
StylePoseGAN is a non-controllable generator to accept conditioning of pose and appearance separately.
Our network can be trained in a fully supervised way with human images to disentangle pose, appearance and body parts.
StylePoseGAN achieves state-of-the-art image generation fidelity on common perceptual metrics.
arXiv Detail & Related papers (2021-02-22T18:50:47Z) - DeepFacePencil: Creating Face Images from Freehand Sketches [77.00929179469559]
Existing image-to-image translation methods require a large-scale dataset of paired sketches and images for supervision.
We propose DeepFacePencil, an effective tool that is able to generate photo-realistic face images from hand-drawn sketches.
arXiv Detail & Related papers (2020-08-31T03:35:21Z) - Deep Generation of Face Images from Sketches [36.146494762987146]
Deep image-to-image translation techniques allow fast generation of face images from freehand sketches.
Existing solutions tend to overfit to sketches, thus requiring professional sketches or even edge maps as input.
We propose to implicitly model the shape space of plausible face images and synthesize a face image in this space to approximate an input sketch.
Our method essentially uses input sketches as soft constraints and is thus able to produce high-quality face images even from rough and/or incomplete sketches.
arXiv Detail & Related papers (2020-06-01T16:20:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.