TAGPerson: A Target-Aware Generation Pipeline for Person
Re-identification
- URL: http://arxiv.org/abs/2112.14239v1
- Date: Tue, 28 Dec 2021 17:56:19 GMT
- Title: TAGPerson: A Target-Aware Generation Pipeline for Person
Re-identification
- Authors: Kai Chen, Weihua Chen, Tao He, Rong Du, Fan Wang, Xiuyu Sun, Yuchen
Guo, Guiguang Ding
- Abstract summary: We propose a novel Target-Aware Generation pipeline to produce synthetic person images, called TAGPerson.
Specifically, it involves a parameterized rendering method, where the parameters are controllable and can be adjusted according to target scenes.
In our experiments, our target-aware synthetic images can achieve a much higher performance than the generalized synthetic images on MSMT17, i.e. 47.5% vs. 40.9% for rank-1 accuracy.
- Score: 65.60874203262375
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Nowadays, real data in person re-identification (ReID) task is facing privacy
issues, e.g., the banned dataset DukeMTMC-ReID. Thus it becomes much harder to
collect real data for ReID task. Meanwhile, the labor cost of labeling ReID
data is still very high and further hinders the development of the ReID
research. Therefore, many methods turn to generate synthetic images for ReID
algorithms as alternatives instead of real images. However, there is an
inevitable domain gap between synthetic and real images. In previous methods,
the generation process is based on virtual scenes, and their synthetic training
data can not be changed according to different target real scenes
automatically. To handle this problem, we propose a novel Target-Aware
Generation pipeline to produce synthetic person images, called TAGPerson.
Specifically, it involves a parameterized rendering method, where the
parameters are controllable and can be adjusted according to target scenes. In
TAGPerson, we extract information from target scenes and use them to control
our parameterized rendering process to generate target-aware synthetic images,
which would hold a smaller gap to the real images in the target domain. In our
experiments, our target-aware synthetic images can achieve a much higher
performance than the generalized synthetic images on MSMT17, i.e. 47.5% vs.
40.9% for rank-1 accuracy. We will release this toolkit\footnote{\noindent Code
is available at
\href{https://github.com/tagperson/tagperson-blender}{https://github.com/tagperson/tagperson-blender}}
for the ReID community to generate synthetic images at any desired taste.
Related papers
- The Unmet Promise of Synthetic Training Images: Using Retrieved Real Images Performs Better [39.57368843211441]
Every synthetic image ultimately originates from the upstream data used to train the generator.
We compare finetuning on task-relevant, targeted synthetic data generated by Stable Diffusion against finetuning on targeted real images retrieved directly from LAION-2B.
Our analysis suggests that this underperformance is partially due to generator artifacts and inaccurate task-relevant visual details in the synthetic images.
arXiv Detail & Related papers (2024-06-07T18:04:21Z) - Bridging Synthetic and Real Worlds for Pre-training Scene Text Detectors [54.80516786370663]
FreeReal is a real-domain-aligned pre-training paradigm that enables the complementary strengths of LSD and real data.
GlyphMix embeds synthetic images as graffiti-like units onto real images.
FreeReal consistently outperforms previous pre-training methods by a substantial margin across four public datasets.
arXiv Detail & Related papers (2023-12-08T15:10:55Z) - FreeMask: Synthetic Images with Dense Annotations Make Stronger
Segmentation Models [62.009002395326384]
FreeMask resorts to synthetic images from generative models to ease the burden of data collection and annotation procedures.
We first synthesize abundant training images conditioned on the semantic masks provided by realistic datasets.
We investigate the role of synthetic images by joint training with real images, or pre-training for real images.
arXiv Detail & Related papers (2023-10-23T17:57:27Z) - Towards Pragmatic Semantic Image Synthesis for Urban Scenes [4.36080478413575]
We present a new task: given a dataset with synthetic images and labels and a dataset with unlabeled real images, our goal is to learn a model that can generate images with the content of the input mask and the appearance of real images.
We leverage the synthetic image as a guide to the content of the generated image by penalizing the difference between their high-level features on a patch level.
In contrast to previous works which employ one discriminator that overfits the target domain semantic distribution, we employ a discriminator for the whole image and multiscale discriminators on the image patches.
arXiv Detail & Related papers (2023-05-16T18:01:12Z) - Is synthetic data from generative models ready for image recognition? [69.42645602062024]
We study whether and how synthetic images generated from state-of-the-art text-to-image generation models can be used for image recognition tasks.
We showcase the powerfulness and shortcomings of synthetic data from existing generative models, and propose strategies for better applying synthetic data for recognition tasks.
arXiv Detail & Related papers (2022-10-14T06:54:24Z) - BLADERUNNER: Rapid Countermeasure for Synthetic (AI-Generated) StyleGAN
Faces [0.0]
StyleGAN is the open-sourced implementation made by NVIDIA.
Report surveys the relevance of AI/ML with respect to Cyber & Information Operations.
Project Blade Runner encompasses two scripts necessary to counter StyleGAN images.
arXiv Detail & Related papers (2022-10-12T21:05:35Z) - A Shared Representation for Photorealistic Driving Simulators [83.5985178314263]
We propose to improve the quality of generated images by rethinking the discriminator architecture.
The focus is on the class of problems where images are generated given semantic inputs, such as scene segmentation maps or human body poses.
We aim to learn a shared latent representation that encodes enough information to jointly do semantic segmentation, content reconstruction, along with a coarse-to-fine grained adversarial reasoning.
arXiv Detail & Related papers (2021-12-09T18:59:21Z) - UnrealPerson: An Adaptive Pipeline towards Costless Person
Re-identification [102.58619642363959]
This paper presents UnrealPerson, a novel pipeline that makes full use of unreal image data to decrease the costs in both the training and deployment stages.
With 3,000 IDs and 120,000 instances, our method achieves a 38.5% rank-1 accuracy when being directly transferred to MSMT17.
arXiv Detail & Related papers (2020-12-08T08:15:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.