Learning Intra-Batch Connections for Deep Metric Learning
- URL: http://arxiv.org/abs/2102.07753v1
- Date: Mon, 15 Feb 2021 18:50:00 GMT
- Title: Learning Intra-Batch Connections for Deep Metric Learning
- Authors: Jenny Seidenschwarz, Ismail Elezi, Laura Leal-Taix\'e
- Abstract summary: metric learning aims to learn a function that maps samples to a lower-dimensional space where similar samples lie closer than dissimilar ones.
Most approaches rely on losses that only take the relations between pairs or triplets of samples into account.
We propose an approach based on message passing networks that takes into account all the relations in a mini-batch.
- Score: 3.5665681694253903
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The goal of metric learning is to learn a function that maps samples to a
lower-dimensional space where similar samples lie closer than dissimilar ones.
In the case of deep metric learning, the mapping is performed by training a
neural network. Most approaches rely on losses that only take the relations
between pairs or triplets of samples into account, which either belong to the
same class or to two different classes. However, these approaches do not
explore the embedding space in its entirety. To this end, we propose an
approach based on message passing networks that takes into account all the
relations in a mini-batch. We refine embedding vectors by exchanging messages
among all samples in a given batch allowing the training process to be aware of
the overall structure. Since not all samples are equally important to predict a
decision boundary, we use dot-product self-attention during message passing to
allow samples to weight the importance of each neighbor accordingly. We achieve
state-of-the-art results on clustering and image retrieval on the CUB-200-2011,
Cars196, Stanford Online Products, and In-Shop Clothes datasets.
Related papers
- Siamese Networks for Weakly Supervised Human Activity Recognition [2.398608007786179]
We present a model with multiple siamese networks that are trained by using only the information about the similarity between pairs of data samples without knowing the explicit labels.
The trained model maps the activity data samples into fixed size representation vectors such that the distance between the vectors in the representation space approximates the similarity of the data samples in the input space.
We evaluate the model on three datasets to verify its effectiveness in segmentation and recognition of continuous human activity sequences.
arXiv Detail & Related papers (2023-07-18T03:23:34Z) - Memorization Through the Lens of Curvature of Loss Function Around
Samples [10.028765645749338]
We propose using the curvature of loss function around each training sample, averaged over training epochs, as a measure of memorization of the sample.
We first show that the high curvature samples visually correspond to long-tailed, mislabeled, or conflicting samples, those that are most likely to be memorized.
This analysis helps us find, to the best of our knowledge, a novel failure mode on the CIFAR100 and ImageNet datasets.
arXiv Detail & Related papers (2023-07-11T22:53:09Z) - Intra-class Adaptive Augmentation with Neighbor Correction for Deep
Metric Learning [99.14132861655223]
We propose a novel intra-class adaptive augmentation (IAA) framework for deep metric learning.
We reasonably estimate intra-class variations for every class and generate adaptive synthetic samples to support hard samples mining.
Our method significantly improves and outperforms the state-of-the-art methods on retrieval performances by 3%-6%.
arXiv Detail & Related papers (2022-11-29T14:52:38Z) - Self-Supervised Arbitrary-Scale Point Clouds Upsampling via Implicit
Neural Representation [79.60988242843437]
We propose a novel approach that achieves self-supervised and magnification-flexible point clouds upsampling simultaneously.
Experimental results demonstrate that our self-supervised learning based scheme achieves competitive or even better performance than supervised learning based state-of-the-art methods.
arXiv Detail & Related papers (2022-04-18T07:18:25Z) - Compare learning: bi-attention network for few-shot learning [6.559037166322981]
One of the Few-shot learning methods called metric learning addresses this challenge by first learning a deep distance metric to determine whether a pair of images belong to the same category.
In this paper, we propose a novel approach named Bi-attention network to compare the instances, which can measure the similarity between embeddings of instances precisely, globally and efficiently.
arXiv Detail & Related papers (2022-03-25T07:39:10Z) - BatchFormer: Learning to Explore Sample Relationships for Robust
Representation Learning [93.38239238988719]
We propose to enable deep neural networks with the ability to learn the sample relationships from each mini-batch.
BatchFormer is applied into the batch dimension of each mini-batch to implicitly explore sample relationships during training.
We perform extensive experiments on over ten datasets and the proposed method achieves significant improvements on different data scarcity applications.
arXiv Detail & Related papers (2022-03-03T05:31:33Z) - Information Bottleneck-Based Hebbian Learning Rule Naturally Ties
Working Memory and Synaptic Updates [0.0]
We take an alternate approach that avoids back-propagation and its associated issues entirely.
Recent work in deep learning proposed independently training each layer of a network via the information bottleneck (IB)
We show that this modulatory signal can be learned by an auxiliary circuit with working memory like a reservoir.
arXiv Detail & Related papers (2021-11-24T17:38:32Z) - Deep Relational Metric Learning [84.95793654872399]
This paper presents a deep relational metric learning framework for image clustering and retrieval.
We learn an ensemble of features that characterizes an image from different aspects to model both interclass and intraclass distributions.
Experiments on the widely-used CUB-200-2011, Cars196, and Stanford Online Products datasets demonstrate that our framework improves existing deep metric learning methods and achieves very competitive results.
arXiv Detail & Related papers (2021-08-23T09:31:18Z) - Exploiting Shared Representations for Personalized Federated Learning [54.65133770989836]
We propose a novel federated learning framework and algorithm for learning a shared data representation across clients and unique local heads for each client.
Our algorithm harnesses the distributed computational power across clients to perform many local-updates with respect to the low-dimensional local parameters for every update of the representation.
This result is of interest beyond federated learning to a broad class of problems in which we aim to learn a shared low-dimensional representation among data distributions.
arXiv Detail & Related papers (2021-02-14T05:36:25Z) - ReMarNet: Conjoint Relation and Margin Learning for Small-Sample Image
Classification [49.87503122462432]
We introduce a novel neural network termed Relation-and-Margin learning Network (ReMarNet)
Our method assembles two networks of different backbones so as to learn the features that can perform excellently in both of the aforementioned two classification mechanisms.
Experiments on four image datasets demonstrate that our approach is effective in learning discriminative features from a small set of labeled samples.
arXiv Detail & Related papers (2020-06-27T13:50:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.