Surpassing Real-World Source Training Data: Random 3D Characters for
Generalizable Person Re-Identification
- URL: http://arxiv.org/abs/2006.12774v2
- Date: Fri, 14 Aug 2020 14:22:24 GMT
- Title: Surpassing Real-World Source Training Data: Random 3D Characters for
Generalizable Person Re-Identification
- Authors: Yanan Wang, Shengcai Liao, Ling Shao
- Abstract summary: We propose to automatically synthesize a large-scale person re-identification dataset following a set-up similar to real surveillance.
We simulate a number of different virtual environments using Unity3D, with customized camera networks similar to real surveillance systems.
As a result, we obtain a virtual dataset, called RandPerson, with 1,801,816 person images of 8,000 identities.
- Score: 109.68210001788506
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Person re-identification has seen significant advancement in recent years.
However, the ability of learned models to generalize to unknown target domains
still remains limited. One possible reason for this is the lack of large-scale
and diverse source training data, since manually labeling such a dataset is
very expensive and privacy sensitive. To address this, we propose to
automatically synthesize a large-scale person re-identification dataset
following a set-up similar to real surveillance but with virtual environments,
and then use the synthesized person images to train a generalizable person
re-identification model. Specifically, we design a method to generate a large
number of random UV texture maps and use them to create different 3D clothing
models. Then, an automatic code is developed to randomly generate various
different 3D characters with diverse clothes, races and attributes. Next, we
simulate a number of different virtual environments using Unity3D, with
customized camera networks similar to real surveillance systems, and import
multiple 3D characters at the same time, with various movements and
interactions along different paths through the camera networks. As a result, we
obtain a virtual dataset, called RandPerson, with 1,801,816 person images of
8,000 identities. By training person re-identification models on these
synthesized person images, we demonstrate, for the first time, that models
trained on virtual data can generalize well to unseen target images, surpassing
the models trained on various real-world datasets, including CUHK03,
Market-1501, DukeMTMC-reID, and almost MSMT17. The RandPerson dataset is
available at https://github.com/VideoObjectSearch/RandPerson.
Related papers
- 3D Human Reconstruction in the Wild with Synthetic Data Using Generative Models [52.96248836582542]
We propose an effective approach based on recent diffusion models, termed HumanWild, which can effortlessly generate human images and corresponding 3D mesh annotations.
By exclusively employing generative models, we generate large-scale in-the-wild human images and high-quality annotations, eliminating the need for real-world data collection.
arXiv Detail & Related papers (2024-03-17T06:31:16Z) - Multi-person 3D pose estimation from unlabelled data [2.54990557236581]
We present a model based on Graph Neural Networks capable of predicting the cross-view correspondence of the people in the scenario.
We also present a Multilayer Perceptron that takes the 2D points to yield the 3D poses of each person.
arXiv Detail & Related papers (2022-12-16T22:03:37Z) - Neural Novel Actor: Learning a Generalized Animatable Neural
Representation for Human Actors [98.24047528960406]
We propose a new method for learning a generalized animatable neural representation from a sparse set of multi-view imagery of multiple persons.
The learned representation can be used to synthesize novel view images of an arbitrary person from a sparse set of cameras, and further animate them with the user's pose control.
arXiv Detail & Related papers (2022-08-25T07:36:46Z) - BEHAVE: Dataset and Method for Tracking Human Object Interactions [105.77368488612704]
We present the first full body human- object interaction dataset with multi-view RGBD frames and corresponding 3D SMPL and object fits along with the annotated contacts between them.
We use this data to learn a model that can jointly track humans and objects in natural environments with an easy-to-use portable multi-camera setup.
arXiv Detail & Related papers (2022-04-14T13:21:19Z) - Cloning Outfits from Real-World Images to 3D Characters for
Generalizable Person Re-Identification [32.85048692231159]
This work proposes an automatic approach to clone the whole outfits from real-world person images to virtual 3D characters.
By rendering the cloned characters in Unity3D scenes, a more realistic virtual dataset called ClonedPerson is created, with 5,621 identities and 887,766 images.
Experimental results show that the model trained on ClonedPerson has a better generalization performance, superior to that trained on other popular real-world and synthetic person re-identification datasets.
arXiv Detail & Related papers (2022-04-06T06:41:08Z) - VoxelTrack: Multi-Person 3D Human Pose Estimation and Tracking in the
Wild [98.69191256693703]
We present VoxelTrack for multi-person 3D pose estimation and tracking from a few cameras which are separated by wide baselines.
It employs a multi-branch network to jointly estimate 3D poses and re-identification (Re-ID) features for all people in the environment.
It outperforms the state-of-the-art methods by a large margin on three public datasets including Shelf, Campus and CMU Panoptic.
arXiv Detail & Related papers (2021-08-05T08:35:44Z) - S3: Neural Shape, Skeleton, and Skinning Fields for 3D Human Modeling [103.65625425020129]
We represent the pedestrian's shape, pose and skinning weights as neural implicit functions that are directly learned from data.
We demonstrate the effectiveness of our approach on various datasets and show that our reconstructions outperform existing state-of-the-art methods.
arXiv Detail & Related papers (2021-01-17T02:16:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.