Cloning Outfits from Real-World Images to 3D Characters for
Generalizable Person Re-Identification
- URL: http://arxiv.org/abs/2204.02611v2
- Date: Thu, 7 Apr 2022 08:39:59 GMT
- Title: Cloning Outfits from Real-World Images to 3D Characters for
Generalizable Person Re-Identification
- Authors: Yanan Wang, Xuezhi Liang, Shengcai Liao
- Abstract summary: This work proposes an automatic approach to clone the whole outfits from real-world person images to virtual 3D characters.
By rendering the cloned characters in Unity3D scenes, a more realistic virtual dataset called ClonedPerson is created, with 5,621 identities and 887,766 images.
Experimental results show that the model trained on ClonedPerson has a better generalization performance, superior to that trained on other popular real-world and synthetic person re-identification datasets.
- Score: 32.85048692231159
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Recently, large-scale synthetic datasets are shown to be very useful for
generalizable person re-identification. However, synthesized persons in
existing datasets are mostly cartoon-like and in random dress collocation,
which limits their performance. To address this, in this work, an automatic
approach is proposed to directly clone the whole outfits from real-world person
images to virtual 3D characters, such that any virtual person thus created will
appear very similar to its real-world counterpart. Specifically, based on UV
texture mapping, two cloning methods are designed, namely registered clothes
mapping and homogeneous cloth expansion. Given clothes keypoints detected on
person images and labeled on regular UV maps with clear clothes structures,
registered mapping applies perspective homography to warp real-world clothes to
the counterparts on the UV map. As for invisible clothes parts and irregular UV
maps, homogeneous expansion segments a homogeneous area on clothes as a
realistic cloth pattern or cell, and expand the cell to fill the UV map.
Furthermore, a similarity-diversity expansion strategy is proposed, by
clustering person images, sampling images per cluster, and cloning outfits for
3D character generation. This way, virtual persons can be scaled up densely in
visual similarity to challenge model learning, and diversely in population to
enrich sample distribution. Finally, by rendering the cloned characters in
Unity3D scenes, a more realistic virtual dataset called ClonedPerson is
created, with 5,621 identities and 887,766 images. Experimental results show
that the model trained on ClonedPerson has a better generalization performance,
superior to that trained on other popular real-world and synthetic person
re-identification datasets. The ClonedPerson project is available at
https://github.com/Yanan-Wang-cs/ClonedPerson.
Related papers
- 3D Human Reconstruction in the Wild with Synthetic Data Using Generative Models [52.96248836582542]
We propose an effective approach based on recent diffusion models, termed HumanWild, which can effortlessly generate human images and corresponding 3D mesh annotations.
By exclusively employing generative models, we generate large-scale in-the-wild human images and high-quality annotations, eliminating the need for real-world data collection.
arXiv Detail & Related papers (2024-03-17T06:31:16Z) - Single-Image 3D Human Digitization with Shape-Guided Diffusion [31.99621159464388]
NeRF and its variants typically require videos or images from different viewpoints.
We present an approach to generate a 360-degree view of a person with a consistent, high-resolution appearance from a single input image.
arXiv Detail & Related papers (2023-11-15T18:59:56Z) - Exploring Shape Embedding for Cloth-Changing Person Re-Identification
via 2D-3D Correspondences [9.487097819140653]
We propose a new shape embedding paradigm for cloth-changing ReID.
The shape embedding paradigm based on 2D-3D correspondences remarkably enhances the model's global understanding of human body shape.
To promote the study of ReID under clothing change, we construct 3D Dense Persons (DP3D), which is the first large-scale cloth-changing ReID dataset.
arXiv Detail & Related papers (2023-10-27T19:26:30Z) - TeCH: Text-guided Reconstruction of Lifelike Clothed Humans [35.68114652041377]
Existing methods often generate overly smooth back-side surfaces with a blurry texture.
Motivated by the power of foundation models, TeCH reconstructs the 3D human by leveraging descriptive text prompts.
We propose a hybrid 3D representation based on DMTet, which consists of an explicit body shape grid and an implicit distance field.
arXiv Detail & Related papers (2023-08-16T17:59:13Z) - One-shot Implicit Animatable Avatars with Model-based Priors [31.385051428938585]
ELICIT is a novel method for learning human-specific neural radiance fields from a single image.
ELICIT has outperformed strong baseline methods of avatar creation when only a single image is available.
arXiv Detail & Related papers (2022-12-05T18:24:06Z) - gDNA: Towards Generative Detailed Neural Avatars [94.9804106939663]
We show that our model is able to generate natural human avatars wearing diverse and detailed clothing.
Our method can be used on the task of fitting human models to raw scans, outperforming the previous state-of-the-art.
arXiv Detail & Related papers (2022-01-11T18:46:38Z) - PISE: Person Image Synthesis and Editing with Decoupled GAN [64.70360318367943]
We propose PISE, a novel two-stage generative model for Person Image Synthesis and Editing.
For human pose transfer, we first synthesize a human parsing map aligned with the target pose to represent the shape of clothing.
To decouple the shape and style of clothing, we propose joint global and local per-region encoding and normalization.
arXiv Detail & Related papers (2021-03-06T04:32:06Z) - Neural 3D Clothes Retargeting from a Single Image [91.5030622330039]
We present a method of clothes; generating the potential poses and deformations of a given 3D clothing template model to fit onto a person in a single RGB image.
The problem is fundamentally ill-posed as attaining the ground truth data is impossible, i.e. images of people wearing the different 3D clothing template model model at exact same pose.
We propose a semi-supervised learning framework that validates the physical plausibility of 3D deformation by matching with the prescribed body-to-cloth contact points and clothing to fit onto the unlabeled silhouette.
arXiv Detail & Related papers (2021-01-29T20:50:34Z) - Surpassing Real-World Source Training Data: Random 3D Characters for
Generalizable Person Re-Identification [109.68210001788506]
We propose to automatically synthesize a large-scale person re-identification dataset following a set-up similar to real surveillance.
We simulate a number of different virtual environments using Unity3D, with customized camera networks similar to real surveillance systems.
As a result, we obtain a virtual dataset, called RandPerson, with 1,801,816 person images of 8,000 identities.
arXiv Detail & Related papers (2020-06-23T05:38:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.