Style-Based Global Appearance Flow for Virtual Try-On
- URL: http://arxiv.org/abs/2204.01046v1
- Date: Sun, 3 Apr 2022 10:58:04 GMT
- Title: Style-Based Global Appearance Flow for Virtual Try-On
- Authors: Sen He, Yi-Zhe Song, Tao Xiang
- Abstract summary: A novel global appearance flow estimation model is proposed in this work.
Experiment results on a popular virtual try-on benchmark show that our method achieves new state-of-the-art performance.
- Score: 119.95115739956661
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Image-based virtual try-on aims to fit an in-shop garment into a clothed
person image. To achieve this, a key step is garment warping which spatially
aligns the target garment with the corresponding body parts in the person
image. Prior methods typically adopt a local appearance flow estimation model.
They are thus intrinsically susceptible to difficult body poses/occlusions and
large mis-alignments between person and garment images (see
Fig.~\ref{fig:fig1}). To overcome this limitation, a novel global appearance
flow estimation model is proposed in this work. For the first time, a StyleGAN
based architecture is adopted for appearance flow estimation. This enables us
to take advantage of a global style vector to encode a whole-image context to
cope with the aforementioned challenges. To guide the StyleGAN flow generator
to pay more attention to local garment deformation, a flow refinement module is
introduced to add local context. Experiment results on a popular virtual try-on
benchmark show that our method achieves new state-of-the-art performance. It is
particularly effective in a `in-the-wild' application scenario where the
reference image is full-body resulting in a large mis-alignment with the
garment image (Fig.~\ref{fig:fig1} Top). Code is available at:
\url{https://github.com/SenHe/Flow-Style-VTON}.
Related papers
- GraVITON: Graph based garment warping with attention guided inversion for Virtual-tryon [5.790630195329777]
We introduce a novel graph based warping technique which emphasizes the value of context in garment flow.
Our method, validated on VITON-HD and Dresscode datasets, showcases substantial improvement in garment warping, texture preservation, and overall realism.
arXiv Detail & Related papers (2024-06-04T10:29:18Z) - MoreStyle: Relax Low-frequency Constraint of Fourier-based Image Reconstruction in Generalizable Medical Image Segmentation [53.24011398381715]
We introduce a Plug-and-Play module for data augmentation called MoreStyle.
MoreStyle diversifies image styles by relaxing low-frequency constraints in Fourier space.
With the help of adversarial learning, MoreStyle pinpoints the most intricate style combinations within latent features.
arXiv Detail & Related papers (2024-03-18T11:38:47Z) - Improving Diffusion Models for Authentic Virtual Try-on in the Wild [53.96244595495942]
This paper considers image-based virtual try-on, which renders an image of a person wearing a curated garment.
We propose a novel diffusion model that improves garment fidelity and generates authentic virtual try-on images.
We present a customization method using a pair of person-garment images, which significantly improves fidelity and authenticity.
arXiv Detail & Related papers (2024-03-08T08:12:18Z) - PG-VTON: A Novel Image-Based Virtual Try-On Method via Progressive
Inference Paradigm [6.929743379017671]
We propose a novel virtual try-on method via progressive inference paradigm (PGVTON)
We exploit the try-on parsing as the shape guidance and implement the garment try-on via warping-mapping-composition.
Experiments demonstrate that our method has state-of-the-art performance under two challenging scenarios.
arXiv Detail & Related papers (2023-04-18T12:47:26Z) - Learning Diverse Tone Styles for Image Retouching [73.60013618215328]
We propose to learn diverse image retouching with normalizing flow-based architectures.
A joint-training pipeline is composed of a style encoder, a conditional RetouchNet, and the image tone style normalizing flow (TSFlow) module.
Our proposed method performs favorably against state-of-the-art methods and is effective in generating diverse results.
arXiv Detail & Related papers (2022-07-12T09:49:21Z) - Controllable Person Image Synthesis with Spatially-Adaptive Warped
Normalization [72.65828901909708]
Controllable person image generation aims to produce realistic human images with desirable attributes.
We introduce a novel Spatially-Adaptive Warped Normalization (SAWN), which integrates a learned flow-field to warp modulation parameters.
We propose a novel self-training part replacement strategy to refine the pretrained model for the texture-transfer task.
arXiv Detail & Related papers (2021-05-31T07:07:44Z) - Towards Photo-Realistic Virtual Try-On by Adaptively
Generating$\leftrightarrow$Preserving Image Content [85.24260811659094]
We propose a novel visual try-on network, namely Adaptive Content Generating and Preserving Network (ACGPN)
ACGPN first predicts semantic layout of the reference image that will be changed after try-on.
Second, a clothes warping module warps clothing images according to the generated semantic layout.
Third, an inpainting module for content fusion integrates all information (e.g. reference image, semantic layout, warped clothes) to adaptively produce each semantic part of human body.
arXiv Detail & Related papers (2020-03-12T15:55:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.