Proxy Anchor Loss for Deep Metric Learning
- URL: http://arxiv.org/abs/2003.13911v1
- Date: Tue, 31 Mar 2020 02:05:27 GMT
- Title: Proxy Anchor Loss for Deep Metric Learning
- Authors: Sungyeon Kim, Dongwon Kim, Minsu Cho, Suha Kwak
- Abstract summary: We present a new proxy-based loss that takes advantages of both pair- and proxy-based methods and overcomes their limitations.
Thanks to the use of proxies, our loss boosts the speed of convergence and is robust against noisy labels and outliers.
Our method is evaluated on four public benchmarks, where a standard network trained with our loss achieves state-of-the-art performance and most quickly converges.
- Score: 47.832107446521626
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Existing metric learning losses can be categorized into two classes:
pair-based and proxy-based losses. The former class can leverage fine-grained
semantic relations between data points, but slows convergence in general due to
its high training complexity. In contrast, the latter class enables fast and
reliable convergence, but cannot consider the rich data-to-data relations. This
paper presents a new proxy-based loss that takes advantages of both pair- and
proxy-based methods and overcomes their limitations. Thanks to the use of
proxies, our loss boosts the speed of convergence and is robust against noisy
labels and outliers. At the same time, it allows embedding vectors of data to
interact with each other in its gradients to exploit data-to-data relations.
Our method is evaluated on four public benchmarks, where a standard network
trained with our loss achieves state-of-the-art performance and most quickly
converges.
Related papers
- JointMatch: A Unified Approach for Diverse and Collaborative
Pseudo-Labeling to Semi-Supervised Text Classification [65.268245109828]
Semi-supervised text classification (SSTC) has gained increasing attention due to its ability to leverage unlabeled data.
Existing approaches based on pseudo-labeling suffer from the issues of pseudo-label bias and error accumulation.
We propose JointMatch, a holistic approach for SSTC that addresses these challenges by unifying ideas from recent semi-supervised learning.
arXiv Detail & Related papers (2023-10-23T05:43:35Z) - Outlier Robust Adversarial Training [57.06824365801612]
We introduce Outlier Robust Adversarial Training (ORAT) in this work.
ORAT is based on a bi-level optimization formulation of adversarial training with a robust rank-based loss function.
We show that the learning objective of ORAT satisfies the $mathcalH$-consistency in binary classification, which establishes it as a proper surrogate to adversarial 0/1 loss.
arXiv Detail & Related papers (2023-09-10T21:36:38Z) - Deep Metric Learning with Soft Orthogonal Proxies [1.823505080809275]
We propose a novel approach that introduces Soft Orthogonality (SO) constraint on proxies.
Our approach leverages Data-Efficient Image Transformer (DeiT) as an encoder to extract contextual features from images along with a DML objective.
Our evaluations demonstrate the superiority of our proposed approach over state-of-the-art methods by a significant margin.
arXiv Detail & Related papers (2023-06-22T17:22:15Z) - SuSana Distancia is all you need: Enforcing class separability in metric
learning via two novel distance-based loss functions for few-shot image
classification [0.9236074230806579]
We propose two loss functions which consider the importance of the embedding vectors by looking at the intra-class and inter-class distance between the few data.
Our results show a significant improvement in accuracy in the miniImagenNet benchmark compared to other metric-based few-shot learning methods by a margin of 2%.
arXiv Detail & Related papers (2023-05-15T23:12:09Z) - Robust Calibrate Proxy Loss for Deep Metric Learning [6.784952050036532]
We propose a Calibrate Proxy structure, which uses the real sample information to improve the similarity calculation in proxy-based loss.
We show that our approach can effectively improve the performance of commonly used proxy-based losses on both regular and noisy datasets.
arXiv Detail & Related papers (2023-04-06T02:43:10Z) - Hierarchical Proxy-based Loss for Deep Metric Learning [32.10423536428467]
Proxy-based metric learning losses are superior to pair-based losses due to their fast convergence and low training complexity.
We present a framework that leverages this implicit hierarchy by imposing a hierarchical structure on the proxies.
Results demonstrate that our hierarchical proxy-based loss framework improves the performance of existing proxy-based losses.
arXiv Detail & Related papers (2021-03-25T00:38:33Z) - Learning from Similarity-Confidence Data [94.94650350944377]
We investigate a novel weakly supervised learning problem of learning from similarity-confidence (Sconf) data.
We propose an unbiased estimator of the classification risk that can be calculated from only Sconf data and show that the estimation error bound achieves the optimal convergence rate.
arXiv Detail & Related papers (2021-02-13T07:31:16Z) - Adaptive additive classification-based loss for deep metric learning [0.0]
We propose an extension to the existing adaptive margin for classification-based deep metric learning.
Our results were achieved with faster convergence and lower code complexity than the prior state-of-the-art.
arXiv Detail & Related papers (2020-06-25T20:45:22Z) - Adversarial Self-Supervised Contrastive Learning [62.17538130778111]
Existing adversarial learning approaches mostly use class labels to generate adversarial samples that lead to incorrect predictions.
We propose a novel adversarial attack for unlabeled data, which makes the model confuse the instance-level identities of the perturbed data samples.
We present a self-supervised contrastive learning framework to adversarially train a robust neural network without labeled data.
arXiv Detail & Related papers (2020-06-13T08:24:33Z) - Towards Certified Robustness of Distance Metric Learning [53.96113074344632]
We advocate imposing an adversarial margin in the input space so as to improve the generalization and robustness of metric learning algorithms.
We show that the enlarged margin is beneficial to the generalization ability by using the theoretical technique of algorithmic robustness.
arXiv Detail & Related papers (2020-06-10T16:51:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.