Strong but Simple Baseline with Dual-Granularity Triplet Loss for
Visible-Thermal Person Re-Identification
- URL: http://arxiv.org/abs/2012.05010v2
- Date: Wed, 10 Mar 2021 08:01:33 GMT
- Title: Strong but Simple Baseline with Dual-Granularity Triplet Loss for
Visible-Thermal Person Re-Identification
- Authors: Haijun Liu, Yanxia Chai, Xiaoheng Tan, Dong Li and Xichuan Zhou
- Abstract summary: We propose a conceptually simple and effective dual-granularity triplet loss for visible-thermal person re-identification (VT-ReID)
Our proposed dual-granularity triplet loss well organizes the sample-based triplet loss and center-based triplet loss in a hierarchical fine to coarse granularity manner.
- Score: 9.964287254346976
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In this letter, we propose a conceptually simple and effective
dual-granularity triplet loss for visible-thermal person re-identification
(VT-ReID). In general, ReID models are always trained with the sample-based
triplet loss and identification loss from the fine granularity level. It is
possible when a center-based loss is introduced to encourage the intra-class
compactness and inter-class discrimination from the coarse granularity level.
Our proposed dual-granularity triplet loss well organizes the sample-based
triplet loss and center-based triplet loss in a hierarchical fine to coarse
granularity manner, just with some simple configurations of typical operations,
such as pooling and batch normalization. Experiments on RegDB and SYSU-MM01
datasets show that with only the global features our dual-granularity triplet
loss can improve the VT-ReID performance by a significant margin. It can be a
strong VT-ReID baseline to boost future research with high quality.
Related papers
- ACTRESS: Active Retraining for Semi-supervised Visual Grounding [52.08834188447851]
A previous study, RefTeacher, makes the first attempt to tackle this task by adopting the teacher-student framework to provide pseudo confidence supervision and attention-based supervision.
This approach is incompatible with current state-of-the-art visual grounding models, which follow the Transformer-based pipeline.
Our paper proposes the ACTive REtraining approach for Semi-Supervised Visual Grounding, abbreviated as ACTRESS.
arXiv Detail & Related papers (2024-07-03T16:33:31Z) - Multi-threshold Deep Metric Learning for Facial Expression Recognition [60.26967776920412]
We present the multi-threshold deep metric learning technique, which avoids the difficult threshold validation.
We find that each threshold of the triplet loss intrinsically determines a distinctive distribution of inter-class variations.
It makes the embedding layer, which is composed of a set of slices, a more informative and discriminative feature.
arXiv Detail & Related papers (2024-06-24T08:27:31Z) - Weak Alignment Supervision from Hybrid Model Improves End-to-end ASR [5.2823268671093775]
We create weak alignment supervision from an existing hybrid system to aid the end-to-end modeling of automatic speech recognition.
We then create a cross-entropy loss at a certain layer of the encoder using the derived alignments.
In contrast to the general one-hot cross-entropy losses, here we use a cross-entropy loss with a label smoothing parameter to regularize the supervision.
arXiv Detail & Related papers (2023-11-24T20:14:28Z) - Learning an Invertible Output Mapping Can Mitigate Simplicity Bias in
Neural Networks [66.76034024335833]
We investigate why diverse/ complex features are learned by the backbone, and their brittleness is due to the linear classification head relying primarily on the simplest features.
We propose Feature Reconstruction Regularizer (FRR) to ensure that the learned features can be reconstructed back from the logits.
We demonstrate up to 15% gains in OOD accuracy on the recently introduced semi-synthetic datasets with extreme distribution shifts.
arXiv Detail & Related papers (2022-10-04T04:01:15Z) - Unified Loss of Pair Similarity Optimization for Vision-Language
Retrieval [35.141916376979836]
There are two popular loss functions used for vision-language retrieval, i.e., triplet loss and contrastive learning loss.
This paper proposes a unified loss of pair similarity optimization for vision-language retrieval.
arXiv Detail & Related papers (2022-09-28T07:01:22Z) - The KFIoU Loss for Rotated Object Detection [115.334070064346]
In this paper, we argue that one effective alternative is to devise an approximate loss who can achieve trend-level alignment with SkewIoU loss.
Specifically, we model the objects as Gaussian distribution and adopt Kalman filter to inherently mimic the mechanism of SkewIoU.
The resulting new loss called KFIoU is easier to implement and works better compared with exact SkewIoU.
arXiv Detail & Related papers (2022-01-29T10:54:57Z) - Label Distributionally Robust Losses for Multi-class Classification:
Consistency, Robustness and Adaptivity [55.29408396918968]
We study a family of loss functions named label-distributionally robust (LDR) losses for multi-class classification.
Our contributions include both consistency and robustness by establishing top-$k$ consistency of LDR losses for multi-class classification.
We propose a new adaptive LDR loss that automatically adapts the individualized temperature parameter to the noise degree of class label of each instance.
arXiv Detail & Related papers (2021-12-30T00:27:30Z) - Deep Ranking with Adaptive Margin Triplet Loss [5.220120772989114]
We propose a simple modification from a fixed margin triplet loss to an adaptive margin triplet loss.
Our proposed loss is well suited for rating datasets in which the ratings are continuous values.
arXiv Detail & Related papers (2021-07-13T15:37:20Z) - Unified Batch All Triplet Loss for Visible-Infrared Person
Re-identification [33.23261883419256]
Batch Hard Triplet loss is widely used in person re-identification tasks, but it does not perform well in the Visible-Infrared person re-identification task.
We propose a batch all triplet selection strategy, which selects all the possible triplets among samples to optimize instead of the hardest triplet.
We also introduce Unified Batch All Triplet loss and Cosine Softmax loss to collaboratively optimize the cosine distance between image vectors.
arXiv Detail & Related papers (2021-03-08T08:58:52Z) - Parameter Sharing Exploration and Hetero-Center based Triplet Loss for
Visible-Thermal Person Re-Identification [17.402673438396345]
This paper focuses on the visible-thermal cross-modality person re-identification (VT Re-ID) task.
Our proposed method distinctly outperforms the state-of-the-art methods by large margins.
arXiv Detail & Related papers (2020-08-14T07:40:35Z) - Beyond Triplet Loss: Meta Prototypical N-tuple Loss for Person
Re-identification [118.72423376789062]
We introduce a multi-class classification loss, i.e., N-tuple loss, to jointly consider multiple (N) instances for per-query optimization.
With the multi-class classification incorporated, our model achieves the state-of-the-art performance on the benchmark person ReID datasets.
arXiv Detail & Related papers (2020-06-08T23:34:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.