GistNet: a Geometric Structure Transfer Network for Long-Tailed
Recognition
- URL: http://arxiv.org/abs/2105.00131v1
- Date: Sat, 1 May 2021 00:37:42 GMT
- Title: GistNet: a Geometric Structure Transfer Network for Long-Tailed
Recognition
- Authors: Bo Liu, Haoxiang Li, Hao Kang, Gang Hua, Nuno Vasconcelos
- Abstract summary: Long-tailed recognition is a problem where the number of examples per class is highly unbalanced.
GistNet is proposed to support this goal, using constellations of classifier parameters to encode the class geometry.
A new learning algorithm is then proposed for GeometrIc Structure Transfer (GIST), with resort to a combination of loss functions that combine class-balanced and random sampling to guarantee that, while overfitting to the popular classes is restricted to geometric parameters, it is leveraged to transfer class geometry from popular to few-shot classes.
- Score: 95.93760490301395
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The problem of long-tailed recognition, where the number of examples per
class is highly unbalanced, is considered. It is hypothesized that the well
known tendency of standard classifier training to overfit to popular classes
can be exploited for effective transfer learning. Rather than eliminating this
overfitting, e.g. by adopting popular class-balanced sampling methods, the
learning algorithm should instead leverage this overfitting to transfer
geometric information from popular to low-shot classes. A new classifier
architecture, GistNet, is proposed to support this goal, using constellations
of classifier parameters to encode the class geometry. A new learning algorithm
is then proposed for GeometrIc Structure Transfer (GIST), with resort to a
combination of loss functions that combine class-balanced and random sampling
to guarantee that, while overfitting to the popular classes is restricted to
geometric parameters, it is leveraged to transfer class geometry from popular
to few-shot classes. This enables better generalization for few-shot classes
without the need for the manual specification of class weights, or even the
explicit grouping of classes into different types. Experiments on two popular
long-tailed recognition datasets show that GistNet outperforms existing
solutions to this problem.
Related papers
- Inductive Graph Few-shot Class Incremental Learning [34.19083477893245]
We introduce inductive GFSCIL that continually learns novel classes with newly emerging nodes.
Compared to the transductive GFSCIL, the inductive setting exacerbates catastrophic forgetting due to inaccessible previous data.
We propose a novel method, called Topology-based class Augmentation and Prototype calibration.
arXiv Detail & Related papers (2024-11-11T00:06:20Z) - Generalized Few-Shot Point Cloud Segmentation Via Geometric Words [54.32239996417363]
Few-shot point cloud segmentation algorithms learn to adapt to new classes at the sacrifice of segmentation accuracy for the base classes.
We present the first attempt at a more practical paradigm of generalized few-shot point cloud segmentation.
We propose the geometric words to represent geometric components shared between the base and novel classes, and incorporate them into a novel geometric-aware semantic representation.
arXiv Detail & Related papers (2023-09-20T11:24:33Z) - Neural Collapse Inspired Feature-Classifier Alignment for Few-Shot Class
Incremental Learning [120.53458753007851]
Few-shot class-incremental learning (FSCIL) has been a challenging problem as only a few training samples are accessible for each novel class in the new sessions.
We deal with this misalignment dilemma in FSCIL inspired by the recently discovered phenomenon named neural collapse.
We propose a neural collapse inspired framework for FSCIL. Experiments on the miniImageNet, CUB-200, and CIFAR-100 datasets demonstrate that our proposed framework outperforms the state-of-the-art performances.
arXiv Detail & Related papers (2023-02-06T18:39:40Z) - Generalization Bounds for Few-Shot Transfer Learning with Pretrained
Classifiers [26.844410679685424]
We study the ability of foundation models to learn representations for classification that are transferable to new, unseen classes.
We show that the few-shot error of the learned feature map on new classes is small in case of class-feature-variability collapse.
arXiv Detail & Related papers (2022-12-23T18:46:05Z) - Evidential Deep Learning for Class-Incremental Semantic Segmentation [15.563703446465823]
Class-Incremental Learning is a challenging problem in machine learning that aims to extend previously trained neural networks with new classes.
In this paper, we address the problem of how to model unlabeled classes while avoiding spurious feature clustering of future uncorrelated classes.
Our method factorizes the problem into a separate foreground class probability, calculated by the expected value of the Dirichlet distribution, and an unknown class (background) probability corresponding to the uncertainty of the estimate.
arXiv Detail & Related papers (2022-12-06T10:13:30Z) - Parametric Classification for Generalized Category Discovery: A Baseline
Study [70.73212959385387]
Generalized Category Discovery (GCD) aims to discover novel categories in unlabelled datasets using knowledge learned from labelled samples.
We investigate the failure of parametric classifiers, verify the effectiveness of previous design choices when high-quality supervision is available, and identify unreliable pseudo-labels as a key problem.
We propose a simple yet effective parametric classification method that benefits from entropy regularisation, achieves state-of-the-art performance on multiple GCD benchmarks and shows strong robustness to unknown class numbers.
arXiv Detail & Related papers (2022-11-21T18:47:11Z) - Multi-Granularity Regularized Re-Balancing for Class Incremental
Learning [32.52884416761171]
Deep learning models suffer from catastrophic forgetting when learning new tasks.
Data imbalance between old and new classes is a key issue that leads to performance degradation of the model.
We propose an assumption-agnostic method, Multi-Granularity Regularized re-Balancing, to address this problem.
arXiv Detail & Related papers (2022-06-30T11:04:51Z) - Do Deep Networks Transfer Invariances Across Classes? [123.84237389985236]
We show how a generative approach for learning the nuisance transformations can help transfer invariances across classes.
Our results provide one explanation for why classifiers generalize poorly on unbalanced and longtailed distributions.
arXiv Detail & Related papers (2022-03-18T04:38:18Z) - Memory-Free Generative Replay For Class-Incremental Learning [32.39857105540859]
We propose a memory-free generative replay strategy to preserve fine-grained old classes characteristics.
Our method is best complemented by prior regularization-based methods proved to be effective for easily distinguishable old classes.
arXiv Detail & Related papers (2021-09-01T12:19:54Z) - Prior Guided Feature Enrichment Network for Few-Shot Segmentation [64.91560451900125]
State-of-the-art semantic segmentation methods require sufficient labeled data to achieve good results.
Few-shot segmentation is proposed to tackle this problem by learning a model that quickly adapts to new classes with a few labeled support samples.
Theses frameworks still face the challenge of generalization ability reduction on unseen classes due to inappropriate use of high-level semantic information.
arXiv Detail & Related papers (2020-08-04T10:41:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.