Equiangular Basis Vectors
- URL: http://arxiv.org/abs/2303.11637v2
- Date: Mon, 8 May 2023 06:25:15 GMT
- Title: Equiangular Basis Vectors
- Authors: Yang Shen and Xuhao Sun and Xiu-Shen Wei
- Abstract summary: In deep neural networks, models usually end with a k-way fully connected layer with softmax to handle different classification tasks.
We propose Equiangular Basis Vectors (EBVs) for classification tasks.
Our EBVs won the first place in the 2022 DIGIX Global AI Challenge.
- Score: 26.520084199562692
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose Equiangular Basis Vectors (EBVs) for classification tasks. In deep
neural networks, models usually end with a k-way fully connected layer with
softmax to handle different classification tasks. The learning objective of
these methods can be summarized as mapping the learned feature representations
to the samples' label space. While in metric learning approaches, the main
objective is to learn a transformation function that maps training data points
from the original space to a new space where similar points are closer while
dissimilar points become farther apart. Different from previous methods, our
EBVs generate normalized vector embeddings as "predefined classifiers" which
are required to not only be with the equal status between each other, but also
be as orthogonal as possible. By minimizing the spherical distance of the
embedding of an input between its categorical EBV in training, the predictions
can be obtained by identifying the categorical EBV with the smallest distance
during inference. Various experiments on the ImageNet-1K dataset and other
downstream tasks demonstrate that our method outperforms the general fully
connected classifier while it does not introduce huge additional computation
compared with classical metric learning methods. Our EBVs won the first place
in the 2022 DIGIX Global AI Challenge, and our code is open-source and
available at https://github.com/NJUST-VIPGroup/Equiangular-Basis-Vectors.
Related papers
- Distance Learner: Incorporating Manifold Prior to Model Training [1.6752182911522522]
We propose a new method, Distance Learner, to incorporate the manifold hypothesis as a prior in modern Deep Neural Networks (DNNs)
For classification, Distance Learner chooses the class corresponding to the closest predicted class manifold.
We evaluate our method on the task of adversarial robustness, and find that it not only outperforms standard classifier by a large margin, but also performs at par with classifiers trained via state-of-the-art adversarial training.
arXiv Detail & Related papers (2022-07-14T13:07:08Z) - Large-Margin Representation Learning for Texture Classification [67.94823375350433]
This paper presents a novel approach combining convolutional layers (CLs) and large-margin metric learning for training supervised models on small datasets for texture classification.
The experimental results on texture and histopathologic image datasets have shown that the proposed approach achieves competitive accuracy with lower computational cost and faster convergence when compared to equivalent CNNs.
arXiv Detail & Related papers (2022-06-17T04:07:45Z) - Compare learning: bi-attention network for few-shot learning [6.559037166322981]
One of the Few-shot learning methods called metric learning addresses this challenge by first learning a deep distance metric to determine whether a pair of images belong to the same category.
In this paper, we propose a novel approach named Bi-attention network to compare the instances, which can measure the similarity between embeddings of instances precisely, globally and efficiently.
arXiv Detail & Related papers (2022-03-25T07:39:10Z) - Do We Really Need a Learnable Classifier at the End of Deep Neural
Network? [118.18554882199676]
We study the potential of learning a neural network for classification with the classifier randomly as an ETF and fixed during training.
Our experimental results show that our method is able to achieve similar performances on image classification for balanced datasets.
arXiv Detail & Related papers (2022-03-17T04:34:28Z) - Transformers Can Do Bayesian Inference [56.99390658880008]
We present Prior-Data Fitted Networks (PFNs)
PFNs leverage in-context learning in large-scale machine learning techniques to approximate a large set of posteriors.
We demonstrate that PFNs can near-perfectly mimic Gaussian processes and also enable efficient Bayesian inference for intractable problems.
arXiv Detail & Related papers (2021-12-20T13:07:39Z) - Rethinking Nearest Neighbors for Visual Classification [56.00783095670361]
k-NN is a lazy learning method that aggregates the distance between the test image and top-k neighbors in a training set.
We adopt k-NN with pre-trained visual representations produced by either supervised or self-supervised methods in two steps.
Via extensive experiments on a wide range of classification tasks, our study reveals the generality and flexibility of k-NN integration.
arXiv Detail & Related papers (2021-12-15T20:15:01Z) - Generalized Learning Vector Quantization for Classification in
Randomized Neural Networks and Hyperdimensional Computing [4.4886210896619945]
We propose a modified RVFL network that avoids computationally expensive matrix operations during training.
The proposed approach achieved state-of-the-art accuracy on a collection of datasets from the UCI Machine Learning Repository.
arXiv Detail & Related papers (2021-06-17T21:17:17Z) - No Fear of Heterogeneity: Classifier Calibration for Federated Learning
with Non-IID Data [78.69828864672978]
A central challenge in training classification models in the real-world federated system is learning with non-IID data.
We propose a novel and simple algorithm called Virtual Representations (CCVR), which adjusts the classifier using virtual representations sampled from an approximated ssian mixture model.
Experimental results demonstrate that CCVR state-of-the-art performance on popular federated learning benchmarks including CIFAR-10, CIFAR-100, and CINIC-10.
arXiv Detail & Related papers (2021-06-09T12:02:29Z) - OVANet: One-vs-All Network for Universal Domain Adaptation [78.86047802107025]
Existing methods manually set a threshold to reject unknown samples based on validation or a pre-defined ratio of unknown samples.
We propose a method to learn the threshold using source samples and to adapt it to the target domain.
Our idea is that a minimum inter-class distance in the source domain should be a good threshold to decide between known or unknown in the target.
arXiv Detail & Related papers (2021-04-07T18:36:31Z) - Few-Shot Incremental Learning with Continually Evolved Classifiers [46.278573301326276]
Few-shot class-incremental learning (FSCIL) aims to design machine learning algorithms that can continually learn new concepts from a few data points.
The difficulty lies in that limited data from new classes not only lead to significant overfitting issues but also exacerbate the notorious catastrophic forgetting problems.
We propose a Continually Evolved CIF ( CEC) that employs a graph model to propagate context information between classifiers for adaptation.
arXiv Detail & Related papers (2021-04-07T10:54:51Z) - Digit Image Recognition Using an Ensemble of One-Versus-All Deep Network
Classifiers [2.385916960125935]
We implement a novel technique for the case of digit image recognition and test and evaluate it on the same.
Every network in an ensemble has been trained by an OVA training technique using the Gradient Descent with Momentum (SGDMA)
Our proposed technique outperforms the baseline on digit image recognition for all datasets.
arXiv Detail & Related papers (2020-06-28T15:37:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.