Connecting Sphere Manifolds Hierarchically for Regularization
- URL: http://arxiv.org/abs/2106.13549v1
- Date: Fri, 25 Jun 2021 10:51:36 GMT
- Title: Connecting Sphere Manifolds Hierarchically for Regularization
- Authors: Damien Scieur, Youngsung Kim
- Abstract summary: We consider classification problems with hierarchically organized classes.
Our technique replaces the last layer of a neural network by combining a spherical fully-connected layer with a hierarchical layer.
This regularization is shown to improve the performance of widely used deep neural network architectures.
- Score: 16.082095595061617
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper considers classification problems with hierarchically organized
classes. We force the classifier (hyperplane) of each class to belong to a
sphere manifold, whose center is the classifier of its super-class. Then,
individual sphere manifolds are connected based on their hierarchical
relations. Our technique replaces the last layer of a neural network by
combining a spherical fully-connected layer with a hierarchical layer. This
regularization is shown to improve the performance of widely used deep neural
network architectures (ResNet and DenseNet) on publicly available datasets
(CIFAR100, CUB200, Stanford dogs, Stanford cars, and Tiny-ImageNet).
Related papers
- CoLaNET -- A Spiking Neural Network with Columnar Layered Architecture for Classification [0.0]
I describe a spiking neural network (SNN) architecture which, can be used in wide range of supervised learning classification tasks.
It is assumed, that all participating signals (the classified object description, correct class label and SNN decision) have spiking nature.
I illustrate the high performance of my network on a task related to model-based reinforcement learning.
arXiv Detail & Related papers (2024-09-02T13:04:54Z) - WLD-Reg: A Data-dependent Within-layer Diversity Regularizer [98.78384185493624]
Neural networks are composed of multiple layers arranged in a hierarchical structure jointly trained with a gradient-based optimization.
We propose to complement this traditional 'between-layer' feedback with additional 'within-layer' feedback to encourage the diversity of the activations within the same layer.
We present an extensive empirical study confirming that the proposed approach enhances the performance of several state-of-the-art neural network models in multiple tasks.
arXiv Detail & Related papers (2023-01-03T20:57:22Z) - A Capsule Network for Hierarchical Multi-Label Image Classification [2.507647327384289]
Hierarchical multi-label classification applies when a multi-class image classification problem is arranged into smaller ones based upon a hierarchy or taxonomy.
We propose a multi-label capsule network (ML-CapsNet) for hierarchical classification.
arXiv Detail & Related papers (2022-09-13T04:17:08Z) - DeepCluE: Enhanced Image Clustering via Multi-layer Ensembles in Deep
Neural Networks [53.88811980967342]
This paper presents a Deep Clustering via Ensembles (DeepCluE) approach.
It bridges the gap between deep clustering and ensemble clustering by harnessing the power of multiple layers in deep neural networks.
Experimental results on six image datasets confirm the advantages of DeepCluE over the state-of-the-art deep clustering approaches.
arXiv Detail & Related papers (2022-06-01T09:51:38Z) - Do We Really Need a Learnable Classifier at the End of Deep Neural
Network? [118.18554882199676]
We study the potential of learning a neural network for classification with the classifier randomly as an ETF and fixed during training.
Our experimental results show that our method is able to achieve similar performances on image classification for balanced datasets.
arXiv Detail & Related papers (2022-03-17T04:34:28Z) - Learning Hierarchical Graph Neural Networks for Image Clustering [81.5841862489509]
We propose a hierarchical graph neural network (GNN) model that learns how to cluster a set of images into an unknown number of identities.
Our hierarchical GNN uses a novel approach to merge connected components predicted at each level of the hierarchy to form a new graph at the next level.
arXiv Detail & Related papers (2021-07-03T01:28:42Z) - Dual-constrained Deep Semi-Supervised Coupled Factorization Network with
Enriched Prior [80.5637175255349]
We propose a new enriched prior based Dual-constrained Deep Semi-Supervised Coupled Factorization Network, called DS2CF-Net.
To ex-tract hidden deep features, DS2CF-Net is modeled as a deep-structure and geometrical structure-constrained neural network.
Our network can obtain state-of-the-art performance for representation learning and clustering.
arXiv Detail & Related papers (2020-09-08T13:10:21Z) - Adaptive Hierarchical Decomposition of Large Deep Networks [4.272649614101117]
As datasets get larger, a natural question is if existing deep learning architectures can be extended to handle the 50+K classes thought to be perceptible by a typical human.
This paper introduces a framework that automatically analyzes and configures a family of smaller deep networks as a replacement to a singular, larger network.
The resulting smaller networks are highly scalable, parallel and more practical to train, and achieve higher classification accuracy.
arXiv Detail & Related papers (2020-07-17T21:04:50Z) - DC-NAS: Divide-and-Conquer Neural Architecture Search [108.57785531758076]
We present a divide-and-conquer (DC) approach to effectively and efficiently search deep neural architectures.
We achieve a $75.1%$ top-1 accuracy on the ImageNet dataset, which is higher than that of state-of-the-art methods using the same search space.
arXiv Detail & Related papers (2020-05-29T09:02:16Z) - Learn Class Hierarchy using Convolutional Neural Networks [0.9569316316728905]
We propose a new architecture for hierarchical classification of images, introducing a stack of deep linear layers with cross-entropy loss functions and center loss combined.
We experimentally show that our hierarchical classifier presents advantages to the traditional classification approaches finding application in computer vision tasks.
arXiv Detail & Related papers (2020-05-18T12:06:43Z) - Do We Need Fully Connected Output Layers in Convolutional Networks? [40.84294968326573]
We show that the typical approach of having a fully connected final output layer is inefficient in terms of parameter count.
We are able to achieve comparable performance to a traditionally learned fully connected classification output layer on the ImageNet-1K, CIFAR-100, Stanford Cars-196, and Oxford Flowers-102 datasets.
arXiv Detail & Related papers (2020-04-28T15:21:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.