NeuralScale: Efficient Scaling of Neurons for Resource-Constrained Deep
Neural Networks
- URL: http://arxiv.org/abs/2006.12813v1
- Date: Tue, 23 Jun 2020 08:14:02 GMT
- Title: NeuralScale: Efficient Scaling of Neurons for Resource-Constrained Deep
Neural Networks
- Authors: Eugene Lee and Chen-Yi Lee
- Abstract summary: We search for the neuron (filter) configuration of a fixed network architecture that maximizes accuracy.
We parameterize the change of the neuron (filter) number of each layer with respect to the change in parameters, allowing us to efficiently scale an architecture across arbitrary sizes.
- Score: 16.518667634574026
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Deciding the amount of neurons during the design of a deep neural network to
maximize performance is not intuitive. In this work, we attempt to search for
the neuron (filter) configuration of a fixed network architecture that
maximizes accuracy. Using iterative pruning methods as a proxy, we parameterize
the change of the neuron (filter) number of each layer with respect to the
change in parameters, allowing us to efficiently scale an architecture across
arbitrary sizes. We also introduce architecture descent which iteratively
refines the parameterized function used for model scaling. The combination of
both proposed methods is coined as NeuralScale. To prove the efficiency of
NeuralScale in terms of parameters, we show empirical simulations on VGG11,
MobileNetV2 and ResNet18 using CIFAR10, CIFAR100 and TinyImageNet as benchmark
datasets. Our results show an increase in accuracy of 3.04%, 8.56% and 3.41%
for VGG11, MobileNetV2 and ResNet18 on CIFAR10, CIFAR100 and TinyImageNet
respectively under a parameter-constrained setting (output neurons (filters) of
default configuration with scaling factor of 0.25).
Related papers
- RelChaNet: Neural Network Feature Selection using Relative Change Scores [0.0]
We introduce RelChaNet, a novel and lightweight feature selection algorithm that uses neuron pruning and regrowth in the input layer of a dense neural network.
Our approach generally outperforms the current state-of-the-art methods, and in particular improves the average accuracy by 2% on the MNIST dataset.
arXiv Detail & Related papers (2024-10-03T09:56:39Z) - Systematic Architectural Design of Scale Transformed Attention Condenser
DNNs via Multi-Scale Class Representational Response Similarity Analysis [93.0013343535411]
We propose a novel type of analysis called Multi-Scale Class Representational Response Similarity Analysis (ClassRepSim)
We show that adding STAC modules to ResNet style architectures can result in up to a 1.6% increase in top-1 accuracy.
Results from ClassRepSim analysis can be used to select an effective parameterization of the STAC module resulting in competitive performance.
arXiv Detail & Related papers (2023-06-16T18:29:26Z) - Receptive Field Refinement for Convolutional Neural Networks Reliably
Improves Predictive Performance [1.52292571922932]
We present a new approach to receptive field analysis that can yield these types of theoretical and empirical performance gains.
Our approach is able to improve ImageNet1K performance across a wide range of well-known, state-of-the-art (SOTA) model classes.
arXiv Detail & Related papers (2022-11-26T05:27:44Z) - EdgeNeXt: Efficiently Amalgamated CNN-Transformer Architecture for
Mobile Vision Applications [68.35683849098105]
We introduce split depth-wise transpose attention (SDTA) encoder that splits input tensors into multiple channel groups.
Our EdgeNeXt model with 1.3M parameters achieves 71.2% top-1 accuracy on ImageNet-1K.
Our EdgeNeXt model with 5.6M parameters achieves 79.4% top-1 accuracy on ImageNet-1K.
arXiv Detail & Related papers (2022-06-21T17:59:56Z) - Variable Bitrate Neural Fields [75.24672452527795]
We present a dictionary method for compressing feature grids, reducing their memory consumption by up to 100x.
We formulate the dictionary optimization as a vector-quantized auto-decoder problem which lets us learn end-to-end discrete neural representations in a space where no direct supervision is available.
arXiv Detail & Related papers (2022-06-15T17:58:34Z) - Neural network relief: a pruning algorithm based on neural activity [47.57448823030151]
We propose a simple importance-score metric that deactivates unimportant connections.
We achieve comparable performance for LeNet architectures on MNIST.
The algorithm is not designed to minimize FLOPs when considering current hardware and software implementations.
arXiv Detail & Related papers (2021-09-22T15:33:49Z) - Recurrent Parameter Generators [42.159272098922685]
We present a generic method for recurrently using the same parameters for many different convolution layers to build a deep network.
We demonstrate how to build a one-layer neural network to achieve similar performance compared to other traditional CNN models.
arXiv Detail & Related papers (2021-07-15T04:23:59Z) - ThriftyNets : Convolutional Neural Networks with Tiny Parameter Budget [4.438259529250529]
We propose a new convolutional neural network architecture, called ThriftyNet.
In ThriftyNet, only one convolutional layer is defined and used recursively, leading to a maximal parameter factorization.
We achieve competitive performance on a tiny parameters budget, exceeding 91% accuracy on CIFAR-10 with less than 40K parameters in total, and 74.3% on CIFAR-100 with less than 600K parameters.
arXiv Detail & Related papers (2020-07-20T13:50:51Z) - When Residual Learning Meets Dense Aggregation: Rethinking the
Aggregation of Deep Neural Networks [57.0502745301132]
We propose Micro-Dense Nets, a novel architecture with global residual learning and local micro-dense aggregations.
Our micro-dense block can be integrated with neural architecture search based models to boost their performance.
arXiv Detail & Related papers (2020-04-19T08:34:52Z) - Widening and Squeezing: Towards Accurate and Efficient QNNs [125.172220129257]
Quantization neural networks (QNNs) are very attractive to the industry because their extremely cheap calculation and storage overhead, but their performance is still worse than that of networks with full-precision parameters.
Most of existing methods aim to enhance performance of QNNs especially binary neural networks by exploiting more effective training techniques.
We address this problem by projecting features in original full-precision networks to high-dimensional quantization features.
arXiv Detail & Related papers (2020-02-03T04:11:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.