CONetV2: Efficient Auto-Channel Size Optimization for CNNs
- URL: http://arxiv.org/abs/2110.06830v1
- Date: Wed, 13 Oct 2021 16:17:19 GMT
- Title: CONetV2: Efficient Auto-Channel Size Optimization for CNNs
- Authors: Yi Ru Wang, Samir Khaki, Weihang Zheng, Mahdi S. Hosseini,
Konstantinos N. Plataniotis
- Abstract summary: This work introduces a method that is efficient in computationally constrained environments by examining the micro-search space of channel size.
In tackling channel-size optimization, we design an automated algorithm to extract the dependencies within different connected layers of the network.
We also introduce a novel metric that highly correlates with test accuracy and enables analysis of individual network layers.
- Score: 35.951376988552695
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Neural Architecture Search (NAS) has been pivotal in finding optimal network
configurations for Convolution Neural Networks (CNNs). While many methods
explore NAS from a global search-space perspective, the employed optimization
schemes typically require heavy computational resources. This work introduces a
method that is efficient in computationally constrained environments by
examining the micro-search space of channel size. In tackling channel-size
optimization, we design an automated algorithm to extract the dependencies
within different connected layers of the network. In addition, we introduce the
idea of knowledge distillation, which enables preservation of trained weights,
admist trials where the channel sizes are changing. Further, since the standard
performance indicators (accuracy, loss) fail to capture the performance of
individual network components (providing an overall network evaluation), we
introduce a novel metric that highly correlates with test accuracy and enables
analysis of individual network layers. Combining dependency extraction,
metrics, and knowledge distillation, we introduce an efficient searching
algorithm, with simulated annealing inspired stochasticity, and demonstrate its
effectiveness in finding optimal architectures that outperform baselines by a
large margin.
Related papers
- Complexity-Aware Training of Deep Neural Networks for Optimal Structure Discovery [0.0]
We propose a novel algorithm for combined unit/filter and layer pruning of deep neural networks that functions during training and without requiring a pre-trained network to apply.
Our algorithm optimally trades-off learning accuracy and pruning levels while balancing layer vs. unit/filter pruning and computational vs. parameter complexity using only three user-defined parameters.
arXiv Detail & Related papers (2024-11-14T02:00:22Z) - Multi-Objective Neural Architecture Search for In-Memory Computing [0.5892638927736115]
We employ neural architecture search (NAS) to enhance the efficiency of deploying diverse machine learning (ML) tasks on in-memory computing architectures.
Our evaluation of this NAS approach for IMC architecture deployment spans three distinct image classification datasets.
arXiv Detail & Related papers (2024-06-10T19:17:09Z) - Pruning-as-Search: Efficient Neural Architecture Search via Channel
Pruning and Structural Reparameterization [50.50023451369742]
Pruning-as-Search (PaS) is an end-to-end channel pruning method to search out desired sub-network automatically and efficiently.
Our proposed architecture outperforms prior arts by around $1.0%$ top-1 accuracy on ImageNet-1000 classification task.
arXiv Detail & Related papers (2022-06-02T17:58:54Z) - CONet: Channel Optimization for Convolutional Neural Networks [33.58529066005248]
We study channel size optimization in convolutional neural networks (CNN)
We introduce an efficient dynamic scaling algorithm -- CONet -- that automatically optimize channel sizes across network layers for a given CNN.
We conduct experiments on CIFAR10/100 and ImageNet datasets and show that CONet can find efficient and accurate architectures searched in ResNet, DARTS, and DARTS+ spaces.
arXiv Detail & Related papers (2021-08-15T21:48:25Z) - Group Fisher Pruning for Practical Network Compression [58.25776612812883]
We present a general channel pruning approach that can be applied to various complicated structures.
We derive a unified metric based on Fisher information to evaluate the importance of a single channel and coupled channels.
Our method can be used to prune any structures including those with coupled channels.
arXiv Detail & Related papers (2021-08-02T08:21:44Z) - Learning Structures for Deep Neural Networks [99.8331363309895]
We propose to adopt the efficient coding principle, rooted in information theory and developed in computational neuroscience.
We show that sparse coding can effectively maximize the entropy of the output signals.
Our experiments on a public image classification dataset demonstrate that using the structure learned from scratch by our proposed algorithm, one can achieve a classification accuracy comparable to the best expert-designed structure.
arXiv Detail & Related papers (2021-05-27T12:27:24Z) - DC-NAS: Divide-and-Conquer Neural Architecture Search [108.57785531758076]
We present a divide-and-conquer (DC) approach to effectively and efficiently search deep neural architectures.
We achieve a $75.1%$ top-1 accuracy on the ImageNet dataset, which is higher than that of state-of-the-art methods using the same search space.
arXiv Detail & Related papers (2020-05-29T09:02:16Z) - AOWS: Adaptive and optimal network width search with latency constraints [30.39613826468697]
We introduce a novel efficient one-shot NAS approach to optimally search for channel numbers.
Experiments on ImageNet classification show that our approach can find networks fitting the resource constraints on different target platforms.
arXiv Detail & Related papers (2020-05-21T06:46:16Z) - Fitting the Search Space of Weight-sharing NAS with Graph Convolutional
Networks [100.14670789581811]
We train a graph convolutional network to fit the performance of sampled sub-networks.
With this strategy, we achieve a higher rank correlation coefficient in the selected set of candidates.
arXiv Detail & Related papers (2020-04-17T19:12:39Z) - Network Adjustment: Channel Search Guided by FLOPs Utilization Ratio [101.84651388520584]
This paper presents a new framework named network adjustment, which considers network accuracy as a function of FLOPs.
Experiments on standard image classification datasets and a wide range of base networks demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2020-04-06T15:51:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.