UnrealPerson: An Adaptive Pipeline towards Costless Person
Re-identification
- URL: http://arxiv.org/abs/2012.04268v2
- Date: Wed, 9 Dec 2020 10:23:56 GMT
- Title: UnrealPerson: An Adaptive Pipeline towards Costless Person
Re-identification
- Authors: Tianyu Zhang and Lingxi Xie and Longhui Wei and Zijie Zhuang and
Yongfei Zhang and Bo Li and Qi Tian
- Abstract summary: This paper presents UnrealPerson, a novel pipeline that makes full use of unreal image data to decrease the costs in both the training and deployment stages.
With 3,000 IDs and 120,000 instances, our method achieves a 38.5% rank-1 accuracy when being directly transferred to MSMT17.
- Score: 102.58619642363959
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The main difficulty of person re-identification (ReID) lies in collecting
annotated data and transferring the model across different domains. This paper
presents UnrealPerson, a novel pipeline that makes full use of unreal image
data to decrease the costs in both the training and deployment stages. Its
fundamental part is a system that can generate synthesized images of
high-quality and from controllable distributions. Instance-level annotation
goes with the synthesized data and is almost free. We point out some details in
image synthesis that largely impact the data quality. With 3,000 IDs and
120,000 instances, our method achieves a 38.5% rank-1 accuracy when being
directly transferred to MSMT17. It almost doubles the former record using
synthesized data and even surpasses previous direct transfer records using real
data. This offers a good basis for unsupervised domain adaption, where our
pre-trained model is easily plugged into the state-of-the-art algorithms
towards higher accuracy. In addition, the data distribution can be flexibly
adjusted to fit some corner ReID scenarios, which widens the application of our
pipeline. We will publish our data synthesis toolkit and synthesized data in
https://github.com/FlyHighest/UnrealPerson.
Related papers
- Let's Synthesize Step by Step: Iterative Dataset Synthesis with Large
Language Models by Extrapolating Errors from Small Models [69.76066070227452]
*Data Synthesis* is a promising way to train a small model with very little labeled data.
We propose *Synthesis Step by Step* (**S3**), a data synthesis framework that shrinks this distribution gap.
Our approach improves the performance of a small model by reducing the gap between the synthetic dataset and the real data.
arXiv Detail & Related papers (2023-10-20T17:14:25Z) - From Fake to Real: Pretraining on Balanced Synthetic Images to Prevent Spurious Correlations in Image Recognition [64.59093444558549]
We propose a simple, easy-to-implement, two-step training pipeline that we call From Fake to Real.
By training on real and synthetic data separately, FFR does not expose the model to the statistical differences between real and synthetic data.
Our experiments show that FFR improves worst group accuracy over the state-of-the-art by up to 20% over three datasets.
arXiv Detail & Related papers (2023-08-08T19:52:28Z) - Synthetic data, real errors: how (not) to publish and use synthetic data [86.65594304109567]
We show how the generative process affects the downstream ML task.
We introduce Deep Generative Ensemble (DGE) to approximate the posterior distribution over the generative process model parameters.
arXiv Detail & Related papers (2023-05-16T07:30:29Z) - A New Benchmark: On the Utility of Synthetic Data with Blender for Bare
Supervised Learning and Downstream Domain Adaptation [42.2398858786125]
Deep learning in computer vision has achieved great success with the price of large-scale labeled training data.
The uncontrollable data collection process produces non-IID training and test data, where undesired duplication may exist.
To circumvent them, an alternative is to generate synthetic data via 3D rendering with domain randomization.
arXiv Detail & Related papers (2023-03-16T09:03:52Z) - Synthesizing Annotated Image and Video Data Using a Rendering-Based
Pipeline for Improved License Plate Recognition [8.15545354580221]
An insufficient number of training samples is a common problem in neural network applications.
We propose a novel rendering-based pipeline for synthesizing annotated data sets.
The pipeline is capable of generating and annotating synthetic and partly-real image and video data in a fully automatic procedure.
arXiv Detail & Related papers (2022-09-28T22:11:58Z) - Virtual passengers for real car solutions: synthetic datasets [2.1028463367241033]
We build a 3D scenario and set-up to resemble reality as closely as possible.
It is possible to configure and vary parameters to add randomness to the scene.
We present the process and concept of synthetic data generation in an automotive context.
arXiv Detail & Related papers (2022-05-13T10:54:39Z) - Self-Supervised Pre-Training for Transformer-Based Person
Re-Identification [54.55281692768765]
Transformer-based supervised pre-training achieves great performance in person re-identification (ReID)
Due to the domain gap between ImageNet and ReID datasets, it usually needs a larger pre-training dataset to boost the performance.
This work aims to mitigate the gap between the pre-training and ReID datasets from the perspective of data and model structure.
arXiv Detail & Related papers (2021-11-23T18:59:08Z) - Fake It Till You Make It: Face analysis in the wild using synthetic data
alone [9.081019005437309]
We show that it is possible to perform face-related computer vision in the wild using synthetic data alone.
We describe how to combine a procedurally-generated 3D face model with a comprehensive library of hand-crafted assets to render training images with unprecedented realism.
arXiv Detail & Related papers (2021-09-30T13:07:04Z) - Contrastive Model Inversion for Data-Free Knowledge Distillation [60.08025054715192]
We propose Contrastive Model Inversion, where the data diversity is explicitly modeled as an optimizable objective.
Our main observation is that, under the constraint of the same amount of data, higher data diversity usually indicates stronger instance discrimination.
Experiments on CIFAR-10, CIFAR-100, and Tiny-ImageNet demonstrate that CMI achieves significantly superior performance when the generated data are used for knowledge distillation.
arXiv Detail & Related papers (2021-05-18T15:13:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.