Maximally Compact and Separated Features with Regular Polytope Networks
- URL: http://arxiv.org/abs/2301.06116v1
- Date: Sun, 15 Jan 2023 15:20:57 GMT
- Title: Maximally Compact and Separated Features with Regular Polytope Networks
- Authors: Federico Pernici, Matteo Bruni, Claudio Baecchi, Alberto Del Bimbo
- Abstract summary: We show how to extract from CNNs features the properties of emphmaximum inter-class separability and emphmaximum intra-class compactness.
We obtain features similar to what can be obtained with the well-known citewen2016discriminative and other similar approaches.
- Score: 22.376196701232388
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Convolutional Neural Networks (CNNs) trained with the Softmax loss are widely
used classification models for several vision tasks. Typically, a learnable
transformation (i.e. the classifier) is placed at the end of such models
returning class scores that are further normalized into probabilities by
Softmax. This learnable transformation has a fundamental role in determining
the network internal feature representation.
In this work we show how to extract from CNNs features with the properties of
\emph{maximum} inter-class separability and \emph{maximum} intra-class
compactness by setting the parameters of the classifier transformation as not
trainable (i.e. fixed). We obtain features similar to what can be obtained with
the well-known ``Center Loss'' \cite{wen2016discriminative} and other similar
approaches but with several practical advantages including maximal exploitation
of the available feature space representation, reduction in the number of
network parameters, no need to use other auxiliary losses besides the Softmax.
Our approach unifies and generalizes into a common approach two apparently
different classes of methods regarding: discriminative features, pioneered by
the Center Loss \cite{wen2016discriminative} and fixed classifiers, firstly
evaluated in \cite{hoffer2018fix}.
Preliminary qualitative experimental results provide some insight on the
potentialities of our combined strategy.
Related papers
- Universal Consistency of Wide and Deep ReLU Neural Networks and Minimax
Optimal Convergence Rates for Kolmogorov-Donoho Optimal Function Classes [7.433327915285969]
We prove the universal consistency of wide and deep ReLU neural network classifiers trained on the logistic loss.
We also give sufficient conditions for a class of probability measures for which classifiers based on neural networks achieve minimax optimal rates of convergence.
arXiv Detail & Related papers (2024-01-08T23:54:46Z) - Variational Classification [51.2541371924591]
We derive a variational objective to train the model, analogous to the evidence lower bound (ELBO) used to train variational auto-encoders.
Treating inputs to the softmax layer as samples of a latent variable, our abstracted perspective reveals a potential inconsistency.
We induce a chosen latent distribution, instead of the implicit assumption found in a standard softmax layer.
arXiv Detail & Related papers (2023-05-17T17:47:19Z) - Equivariance with Learned Canonicalization Functions [77.32483958400282]
We show that learning a small neural network to perform canonicalization is better than using predefineds.
Our experiments show that learning the canonicalization function is competitive with existing techniques for learning equivariant functions across many tasks.
arXiv Detail & Related papers (2022-11-11T21:58:15Z) - Do We Really Need a Learnable Classifier at the End of Deep Neural
Network? [118.18554882199676]
We study the potential of learning a neural network for classification with the classifier randomly as an ETF and fixed during training.
Our experimental results show that our method is able to achieve similar performances on image classification for balanced datasets.
arXiv Detail & Related papers (2022-03-17T04:34:28Z) - Exploring Category-correlated Feature for Few-shot Image Classification [27.13708881431794]
We present a simple yet effective feature rectification method by exploring the category correlation between novel and base classes as the prior knowledge.
The proposed approach consistently obtains considerable performance gains on three widely used benchmarks.
arXiv Detail & Related papers (2021-12-14T08:25:24Z) - Learning Debiased and Disentangled Representations for Semantic
Segmentation [52.35766945827972]
We propose a model-agnostic and training scheme for semantic segmentation.
By randomly eliminating certain class information in each training iteration, we effectively reduce feature dependencies among classes.
Models trained with our approach demonstrate strong results on multiple semantic segmentation benchmarks.
arXiv Detail & Related papers (2021-10-31T16:15:09Z) - Regular Polytope Networks [29.44144177954405]
We argue that a transformation can be fixed with no loss of accuracy and with a reduction in memory usage.
It can also be used to learn stationary and maximally separated embeddings.
We show that the stationarity of the embedding and its maximal separated representation can be theoretically justified.
arXiv Detail & Related papers (2021-03-29T14:11:32Z) - Hyperspherical embedding for novel class classification [1.5952956981784217]
We present a constraint-based approach applied to representations in the latent space under the normalized softmax loss.
We experimentally validate the proposed approach for the classification of unseen classes on different datasets using both metric learning and the normalized softmax loss.
Our results show that not only our proposed strategy can be efficiently trained on larger set of classes, as it does not require pairwise learning, but also present better classification results than the metric learning strategies.
arXiv Detail & Related papers (2021-02-05T15:42:13Z) - Shaping Deep Feature Space towards Gaussian Mixture for Visual
Classification [74.48695037007306]
We propose a Gaussian mixture (GM) loss function for deep neural networks for visual classification.
With a classification margin and a likelihood regularization, the GM loss facilitates both high classification performance and accurate modeling of the feature distribution.
The proposed model can be implemented easily and efficiently without using extra trainable parameters.
arXiv Detail & Related papers (2020-11-18T03:32:27Z) - ReMarNet: Conjoint Relation and Margin Learning for Small-Sample Image
Classification [49.87503122462432]
We introduce a novel neural network termed Relation-and-Margin learning Network (ReMarNet)
Our method assembles two networks of different backbones so as to learn the features that can perform excellently in both of the aforementioned two classification mechanisms.
Experiments on four image datasets demonstrate that our approach is effective in learning discriminative features from a small set of labeled samples.
arXiv Detail & Related papers (2020-06-27T13:50:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.