A High-Accuracy Unsupervised Person Re-identification Method Using
Auxiliary Information Mined from Datasets
- URL: http://arxiv.org/abs/2205.03124v1
- Date: Fri, 6 May 2022 10:16:18 GMT
- Title: A High-Accuracy Unsupervised Person Re-identification Method Using
Auxiliary Information Mined from Datasets
- Authors: Hehan Teng, Tao He, Yuchen Guo, Guiguang Ding
- Abstract summary: We make use of auxiliary information mined from datasets for multi-modal feature learning.
This paper proposes three effective training tricks, including Restricted Label Smoothing Cross Entropy Loss (RLSCE), Weight Adaptive Triplet Loss (WATL) and Dynamic Training Iterations (DTI)
- Score: 53.047542904329866
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Supervised person re-identification methods rely heavily on high-quality
cross-camera training label. This significantly hinders the deployment of re-ID
models in real-world applications. The unsupervised person re-ID methods can
reduce the cost of data annotation, but their performance is still far lower
than the supervised ones. In this paper, we make full use of the auxiliary
information mined from the datasets for multi-modal feature learning, including
camera information, temporal information and spatial information. By analyzing
the style bias of cameras, the characteristics of pedestrians' motion
trajectories and the positions of camera network, this paper designs three
modules: Time-Overlapping Constraint (TOC), Spatio-Temporal Similarity (STS)
and Same-Camera Penalty (SCP) to exploit the auxiliary information. Auxiliary
information can improve the model performance and inference accuracy by
constructing association constraints or fusing with visual features. In
addition, this paper proposes three effective training tricks, including
Restricted Label Smoothing Cross Entropy Loss (RLSCE), Weight Adaptive Triplet
Loss (WATL) and Dynamic Training Iterations (DTI). The tricks achieve mAP of
72.4% and 81.1% on MARS and DukeMTMC-VideoReID, respectively. Combined with
auxiliary information exploiting modules, our methods achieve mAP of 89.9% on
DukeMTMC, where TOC, STS and SCP all contributed considerable performance
improvements. The method proposed by this paper outperforms most existing
unsupervised re-ID methods and narrows the gap between unsupervised and
supervised re-ID methods. Our code is at
https://github.com/tenghehan/AuxUSLReID.
Related papers
- DVPE: Divided View Position Embedding for Multi-View 3D Object Detection [7.791229698270439]
Current research faces challenges in balancing between receptive fields and reducing interference when aggregating multi-view features.
This paper proposes a divided view method, in which features are modeled globally via the visibility crossattention mechanism, but interact only with partial features in a divided local virtual space.
Our framework, named DVPE, achieves state-of-the-art performance (57.2% mAP and 64.5% NDS) on the nuScenes test set.
arXiv Detail & Related papers (2024-07-24T02:44:41Z) - Benchmarking the Robustness of LiDAR-Camera Fusion for 3D Object
Detection [58.81316192862618]
Two critical sensors for 3D perception in autonomous driving are the camera and the LiDAR.
fusing these two modalities can significantly boost the performance of 3D perception models.
We benchmark the state-of-the-art fusion methods for the first time.
arXiv Detail & Related papers (2022-05-30T09:35:37Z) - Camera-Tracklet-Aware Contrastive Learning for Unsupervised Vehicle
Re-Identification [4.5471611558189124]
We propose camera-tracklet-aware contrastive learning (CTACL) using the multi-camera tracklet information without vehicle identity labels.
The proposed CTACL divides an unlabelled domain, i.e., entire vehicle images, into multiple camera-level images and conducts contrastive learning.
We demonstrate the effectiveness of our approach on video-based and image-based vehicle Re-ID datasets.
arXiv Detail & Related papers (2021-09-14T02:12:54Z) - Unsupervised Noisy Tracklet Person Re-identification [100.85530419892333]
We present a novel selective tracklet learning (STL) approach that can train discriminative person re-id models from unlabelled tracklet data.
This avoids the tedious and costly process of exhaustively labelling person image/tracklet true matching pairs across camera views.
Our method is particularly more robust against arbitrary noisy data of raw tracklets therefore scalable to learning discriminative models from unconstrained tracking data.
arXiv Detail & Related papers (2021-01-16T07:31:00Z) - Camera-aware Proxies for Unsupervised Person Re-Identification [60.26031011794513]
This paper tackles the purely unsupervised person re-identification (Re-ID) problem that requires no annotations.
We propose to split each single cluster into multiple proxies and each proxy represents the instances coming from the same camera.
Based on the camera-aware proxies, we design both intra- and inter-camera contrastive learning components for our Re-ID model.
arXiv Detail & Related papers (2020-12-19T12:37:04Z) - Attribute-aware Identity-hard Triplet Loss for Video-based Person
Re-identification [51.110453988705395]
Video-based person re-identification (Re-ID) is an important computer vision task.
We introduce a new metric learning method called Attribute-aware Identity-hard Triplet Loss (AITL)
To achieve a complete model of video-based person Re-ID, a multi-task framework with Attribute-driven Spatio-Temporal Attention (ASTA) mechanism is also proposed.
arXiv Detail & Related papers (2020-06-13T09:15:38Z) - Intra-Camera Supervised Person Re-Identification [87.88852321309433]
We propose a novel person re-identification paradigm based on an idea of independent per-camera identity annotation.
This eliminates the most time-consuming and tedious inter-camera identity labelling process.
We formulate a Multi-tAsk mulTi-labEl (MATE) deep learning method for Intra-Camera Supervised (ICS) person re-id.
arXiv Detail & Related papers (2020-02-12T15:26:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.