Memory-Augmented Relation Network for Few-Shot Learning
- URL: http://arxiv.org/abs/2005.04414v2
- Date: Wed, 13 May 2020 03:25:36 GMT
- Title: Memory-Augmented Relation Network for Few-Shot Learning
- Authors: Jun He, Richang Hong, Xueliang Liu, Mingliang Xu, Zhengjun Zha and
Meng Wang
- Abstract summary: In this work, we investigate a new metric-learning method, Memory-Augmented Relation Network (MRN)
In MRN, we choose the samples that are visually similar from the working context, and perform weighted information propagation to attentively aggregate helpful information from chosen ones to enhance its representation.
We empirically demonstrate that MRN yields significant improvement over its ancestor and achieves competitive or even better performance when compared with other few-shot learning approaches.
- Score: 114.47866281436829
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Metric-based few-shot learning methods concentrate on learning transferable
feature embedding that generalizes well from seen categories to unseen
categories under the supervision of limited number of labelled instances.
However, most of them treat each individual instance in the working context
separately without considering its relationships with the others. In this work,
we investigate a new metric-learning method, Memory-Augmented Relation Network
(MRN), to explicitly exploit these relationships. In particular, for an
instance, we choose the samples that are visually similar from the working
context, and perform weighted information propagation to attentively aggregate
helpful information from the chosen ones to enhance its representation. In MRN,
we also formulate the distance metric as a learnable relation module which
learns to compare for similarity measurement, and augment the working context
with memory slots, both contributing to its generality. We empirically
demonstrate that MRN yields significant improvement over its ancestor and
achieves competitive or even better performance when compared with other
few-shot learning approaches on the two major benchmark datasets, i.e.
miniImagenet and tieredImagenet.
Related papers
- Metric Learning as a Service with Covariance Embedding [7.5989847759545155]
Metric learning aims to maximize and minimize inter- and intra-class similarities.
Existing models mainly rely on distance measures to obtain a separable embedding space.
We argue that to enable metric learning as a service for high-performance deep learning applications, we should also wisely deal with inter-class relationships.
arXiv Detail & Related papers (2022-11-28T10:10:59Z) - Weak Augmentation Guided Relational Self-Supervised Learning [80.0680103295137]
We introduce a novel relational self-supervised learning (ReSSL) framework that learns representations by modeling the relationship between different instances.
Our proposed method employs sharpened distribution of pairwise similarities among different instances as textitrelation metric.
Experimental results show that our proposed ReSSL substantially outperforms the state-of-the-art methods across different network architectures.
arXiv Detail & Related papers (2022-03-16T16:14:19Z) - Deep Relational Metric Learning [84.95793654872399]
This paper presents a deep relational metric learning framework for image clustering and retrieval.
We learn an ensemble of features that characterizes an image from different aspects to model both interclass and intraclass distributions.
Experiments on the widely-used CUB-200-2011, Cars196, and Stanford Online Products datasets demonstrate that our framework improves existing deep metric learning methods and achieves very competitive results.
arXiv Detail & Related papers (2021-08-23T09:31:18Z) - ReSSL: Relational Self-Supervised Learning with Weak Augmentation [68.47096022526927]
Self-supervised learning has achieved great success in learning visual representations without data annotations.
We introduce a novel relational SSL paradigm that learns representations by modeling the relationship between different instances.
Our proposed ReSSL significantly outperforms the previous state-of-the-art algorithms in terms of both performance and training efficiency.
arXiv Detail & Related papers (2021-07-20T06:53:07Z) - Revisiting Contrastive Methods for Unsupervised Learning of Visual
Representations [78.12377360145078]
Contrastive self-supervised learning has outperformed supervised pretraining on many downstream tasks like segmentation and object detection.
In this paper, we first study how biases in the dataset affect existing methods.
We show that current contrastive approaches work surprisingly well across: (i) object- versus scene-centric, (ii) uniform versus long-tailed and (iii) general versus domain-specific datasets.
arXiv Detail & Related papers (2021-06-10T17:59:13Z) - ATRM: Attention-based Task-level Relation Module for GNN-based Few-shot
Learning [14.464964336101028]
We propose a new relation measure method, namely the attention-based task-level relation module (ATRM)
The proposed module captures the relation representations between nodes by considering the sample-to-task instead of sample-to-sample embedding features.
Experimental results demonstrate that the proposed module is effective for GNN-based few-shot learning.
arXiv Detail & Related papers (2021-01-25T00:53:04Z) - Beyond Single Instance Multi-view Unsupervised Representation Learning [21.449132256091662]
We impose more accurate instance discrimination capability by measuring the joint similarity between two randomly sampled instances.
We believe that learning joint similarity helps to improve the performance when encoded features are distributed more evenly in the latent space.
arXiv Detail & Related papers (2020-11-26T15:43:27Z) - Region Comparison Network for Interpretable Few-shot Image
Classification [97.97902360117368]
Few-shot image classification has been proposed to effectively use only a limited number of labeled examples to train models for new classes.
We propose a metric learning based method named Region Comparison Network (RCN), which is able to reveal how few-shot learning works.
We also present a new way to generalize the interpretability from the level of tasks to categories.
arXiv Detail & Related papers (2020-09-08T07:29:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.