Graph Sampling Based Deep Metric Learning for Generalizable Person
Re-Identification
- URL: http://arxiv.org/abs/2104.01546v2
- Date: Tue, 6 Apr 2021 05:26:26 GMT
- Title: Graph Sampling Based Deep Metric Learning for Generalizable Person
Re-Identification
- Authors: Shengcai Liao and Ling Shao
- Abstract summary: We argue that the most popular random sampling method, the well-known PK sampler, is not informative and efficient for deep metric learning.
We propose an efficient mini batch sampling method called Graph Sampling (GS) for large-scale metric learning.
- Score: 114.56752624945142
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Generalizable person re-identification has recently got increasing attention
due to its research values as well as practical values. However, the efficiency
of learning from large-scale data has not yet been much studied. In this paper,
we argue that the most popular random sampling method, the well-known PK
sampler, is not informative and efficient for deep metric learning. Though
online hard example mining improves the learning efficiency to some extent, the
mining in mini batches after random sampling is still limited. Therefore, this
inspires us that the hard example mining should be shifted backward to the data
sampling stage. To address this, in this paper, we propose an efficient mini
batch sampling method called Graph Sampling (GS) for large-scale metric
learning. The basic idea is to build a nearest neighbor relationship graph for
all classes at the beginning of each epoch. Then, each mini batch is composed
of a randomly selected class and its nearest neighboring classes so as to
provide informative and challenging examples for learning. Together with an
adapted competitive baseline, we improve the previous state of the arts in
generalizable person re-identification significantly, by up to 22.3% in Rank-1
and 15% in mAP. Besides, the proposed method also outperforms the competitive
baseline by up to 4%, with the training time significantly reduced by up to
x6.6, from 12.2 hours to 1.8 hours in training a large-scale dataset RandPerson
with 8,000 IDs. Code is available at
\url{https://github.com/ShengcaiLiao/QAConv}.
Related papers
- KAKURENBO: Adaptively Hiding Samples in Deep Neural Network Training [2.8804804517897935]
We propose a method for hiding the least-important samples during the training of deep neural networks.
We adaptively find samples to exclude in a given epoch based on their contribution to the overall learning process.
Our method can reduce total training time by up to 22% impacting accuracy only by 0.4% compared to the baseline.
arXiv Detail & Related papers (2023-10-16T06:19:29Z) - Global Proxy-based Hard Mining for Visual Place Recognition [3.6739949215165164]
We introduce a new technique that performs global hard mini-batch sampling based on proxies.
To do so, we add a new end-to-end trainable branch to the network, which generates efficient place descriptors.
Our method can be used in combination with all existing pairwise and triplet loss functions with negligible additional memory and computation cost.
arXiv Detail & Related papers (2023-02-28T00:43:13Z) - Intra-class Adaptive Augmentation with Neighbor Correction for Deep
Metric Learning [99.14132861655223]
We propose a novel intra-class adaptive augmentation (IAA) framework for deep metric learning.
We reasonably estimate intra-class variations for every class and generate adaptive synthetic samples to support hard samples mining.
Our method significantly improves and outperforms the state-of-the-art methods on retrieval performances by 3%-6%.
arXiv Detail & Related papers (2022-11-29T14:52:38Z) - A Maximum Log-Likelihood Method for Imbalanced Few-Shot Learning Tasks [3.2895195535353308]
We propose a new maximum log-likelihood metric for few-shot architectures.
We demonstrate that the proposed metric achieves superior performance accuracy w.r.t. conventional similarity metrics.
We also show that our algorithm achieves state-of-the-art transductive few-shot performance when the evaluation data is imbalanced.
arXiv Detail & Related papers (2022-11-26T21:31:00Z) - Global Contrastive Batch Sampling via Optimization on Sample
Permutations [28.72288652451881]
Global Contrastive Batch Sampling (GCBS) is an efficient approximation to the batch assignment problem.
GCBS improves state-of-the-art performance in sentence embedding and code-search tasks.
arXiv Detail & Related papers (2022-10-23T22:35:02Z) - Learning Fast Sample Re-weighting Without Reward Data [41.92662851886547]
This paper presents a novel learning-based fast sample re-weighting (FSR) method that does not require additional reward data.
Our experiments show the proposed method achieves competitive results compared to state of the arts on label noise and long-tailed recognition.
arXiv Detail & Related papers (2021-09-07T17:30:56Z) - Attentional-Biased Stochastic Gradient Descent [74.49926199036481]
We present a provable method (named ABSGD) for addressing the data imbalance or label noise problem in deep learning.
Our method is a simple modification to momentum SGD where we assign an individual importance weight to each sample in the mini-batch.
ABSGD is flexible enough to combine with other robust losses without any additional cost.
arXiv Detail & Related papers (2020-12-13T03:41:52Z) - Fewer is More: A Deep Graph Metric Learning Perspective Using Fewer
Proxies [65.92826041406802]
We propose a Proxy-based deep Graph Metric Learning approach from the perspective of graph classification.
Multiple global proxies are leveraged to collectively approximate the original data points for each class.
We design a novel reverse label propagation algorithm, by which the neighbor relationships are adjusted according to ground-truth labels.
arXiv Detail & Related papers (2020-10-26T14:52:42Z) - Learning to Count in the Crowd from Limited Labeled Data [109.2954525909007]
We focus on reducing the annotation efforts by learning to count in the crowd from limited number of labeled samples.
Specifically, we propose a Gaussian Process-based iterative learning mechanism that involves estimation of pseudo-ground truth for the unlabeled data.
arXiv Detail & Related papers (2020-07-07T04:17:01Z) - Omni-supervised Facial Expression Recognition via Distilled Data [120.11782405714234]
We propose omni-supervised learning to exploit reliable samples in a large amount of unlabeled data for network training.
We experimentally verify that the new dataset can significantly improve the ability of the learned FER model.
To tackle this, we propose to apply a dataset distillation strategy to compress the created dataset into several informative class-wise images.
arXiv Detail & Related papers (2020-05-18T09:36:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.