Many-Class Few-Shot Learning on Multi-Granularity Class Hierarchy
- URL: http://arxiv.org/abs/2006.15479v1
- Date: Sun, 28 Jun 2020 01:11:34 GMT
- Title: Many-Class Few-Shot Learning on Multi-Granularity Class Hierarchy
- Authors: Lu Liu, Tianyi Zhou, Guodong Long, Jing Jiang and Chengqi Zhang
- Abstract summary: We study many-class few-shot (MCFS) problem in both supervised learning and meta-learning settings.
In this paper, we leverage the class hierarchy as a prior knowledge to train a coarse-to-fine classifier.
The model, "memory-augmented hierarchical-classification network (MahiNet)", performs coarse-to-fine classification where each coarse class can cover multiple fine classes.
- Score: 57.68486382473194
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study many-class few-shot (MCFS) problem in both supervised learning and
meta-learning settings. Compared to the well-studied many-class many-shot and
few-class few-shot problems, the MCFS problem commonly occurs in practical
applications but has been rarely studied in previous literature. It brings new
challenges of distinguishing between many classes given only a few training
samples per class. In this paper, we leverage the class hierarchy as a prior
knowledge to train a coarse-to-fine classifier that can produce accurate
predictions for MCFS problem in both settings. The propose model,
"memory-augmented hierarchical-classification network (MahiNet)", performs
coarse-to-fine classification where each coarse class can cover multiple fine
classes. Since it is challenging to directly distinguish a variety of fine
classes given few-shot data per class, MahiNet starts from learning a
classifier over coarse-classes with more training data whose labels are much
cheaper to obtain. The coarse classifier reduces the searching range over the
fine classes and thus alleviates the challenges from "many classes". On
architecture, MahiNet firstly deploys a convolutional neural network (CNN) to
extract features. It then integrates a memory-augmented attention module and a
multi-layer perceptron (MLP) together to produce the probabilities over coarse
and fine classes. While the MLP extends the linear classifier, the attention
module extends the KNN classifier, both together targeting the "few-shot"
problem. We design several training strategies of MahiNet for supervised
learning and meta-learning. In addition, we propose two novel benchmark
datasets "mcfsImageNet" and "mcfsOmniglot" specially designed for MCFS problem.
In experiments, we show that MahiNet outperforms several state-of-the-art
models on MCFS problems in both supervised learning and meta-learning.
Related papers
- Knowledge Adaptation Network for Few-Shot Class-Incremental Learning [23.90555521006653]
Few-shot class-incremental learning aims to incrementally recognize new classes using a few samples.
One of the effective methods to solve this challenge is to construct prototypical evolution classifiers.
Because representations for new classes are weak and biased, we argue such a strategy is suboptimal.
arXiv Detail & Related papers (2024-09-18T07:51:38Z) - How to Train Your MAML to Excel in Few-Shot Classification [26.51244463209443]
We show how to train MAML to excel in few-shot classification.
Our approach, which we name UNICORN-MAML, performs on a par with or even outperforms state-of-the-art algorithms.
arXiv Detail & Related papers (2021-06-30T17:56:15Z) - Multi-Class Classification from Single-Class Data with Confidences [90.48669386745361]
We propose an empirical risk minimization framework that is loss-/model-/optimizer-independent.
We show that our method can be Bayes-consistent with a simple modification even if the provided confidences are highly noisy.
arXiv Detail & Related papers (2021-06-16T15:38:13Z) - Dynamic Class Queue for Large Scale Face Recognition In the Wild [45.3063075576461]
This work focus on computing resource constraint and long-tailed class distribution.
We propose a dynamic class queue (DCQ) to tackle these two problems.
We empirically verify in large-scale datasets that 10% of classes are sufficient to achieve similar performance as using all classes.
arXiv Detail & Related papers (2021-05-24T06:31:10Z) - GistNet: a Geometric Structure Transfer Network for Long-Tailed
Recognition [95.93760490301395]
Long-tailed recognition is a problem where the number of examples per class is highly unbalanced.
GistNet is proposed to support this goal, using constellations of classifier parameters to encode the class geometry.
A new learning algorithm is then proposed for GeometrIc Structure Transfer (GIST), with resort to a combination of loss functions that combine class-balanced and random sampling to guarantee that, while overfitting to the popular classes is restricted to geometric parameters, it is leveraged to transfer class geometry from popular to few-shot classes.
arXiv Detail & Related papers (2021-05-01T00:37:42Z) - Improving Calibration for Long-Tailed Recognition [68.32848696795519]
We propose two methods to improve calibration and performance in such scenarios.
For dataset bias due to different samplers, we propose shifted batch normalization.
Our proposed methods set new records on multiple popular long-tailed recognition benchmark datasets.
arXiv Detail & Related papers (2021-04-01T13:55:21Z) - Fine-grained Angular Contrastive Learning with Coarse Labels [72.80126601230447]
We introduce a novel 'Angular normalization' module that allows to effectively combine supervised and self-supervised contrastive pre-training.
This work will help to pave the way for future research on this new, challenging, and very practical topic of C2FS classification.
arXiv Detail & Related papers (2020-12-07T08:09:02Z) - Learning Adaptive Embedding Considering Incremental Class [55.21855842960139]
Class-Incremental Learning (CIL) aims to train a reliable model with the streaming data, which emerges unknown classes sequentially.
Different from traditional closed set learning, CIL has two main challenges: 1) Novel class detection.
After the novel classes are detected, the model needs to be updated without re-training using entire previous data.
arXiv Detail & Related papers (2020-08-31T04:11:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.