Neural Inheritance Relation Guided One-Shot Layer Assignment Search
- URL: http://arxiv.org/abs/2002.12580v1
- Date: Fri, 28 Feb 2020 07:40:48 GMT
- Title: Neural Inheritance Relation Guided One-Shot Layer Assignment Search
- Authors: Rang Meng, Weijie Chen, Di Xie, Yuan Zhang, Shiliang Pu
- Abstract summary: We investigate the impact of different layer assignments to the network performance by building an architecture dataset of layer assignment on CIFAR-100.
We find a neural inheritance relation among the networks with different layer assignments, that is, the optimal layer assignments for deeper networks always inherit from those for shallow networks.
Inspired by this neural inheritance relation, we propose an efficient one-shot layer assignment search approach via inherited sampling.
- Score: 44.82474044430184
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Layer assignment is seldom picked out as an independent research topic in
neural architecture search. In this paper, for the first time, we
systematically investigate the impact of different layer assignments to the
network performance by building an architecture dataset of layer assignment on
CIFAR-100. Through analyzing this dataset, we discover a neural inheritance
relation among the networks with different layer assignments, that is, the
optimal layer assignments for deeper networks always inherit from those for
shallow networks. Inspired by this neural inheritance relation, we propose an
efficient one-shot layer assignment search approach via inherited sampling.
Specifically, the optimal layer assignment searched in the shallow network can
be provided as a strong sampling priori to train and search the deeper ones in
supernet, which extremely reduces the network search space. Comprehensive
experiments carried out on CIFAR-100 illustrate the efficiency of our proposed
method. Our search results are strongly consistent with the optimal ones
directly selected from the architecture dataset. To further confirm the
generalization of our proposed method, we also conduct experiments on
Tiny-ImageNet and ImageNet. Our searched results are remarkably superior to the
handcrafted ones under the unchanged computational budgets. The neural
inheritance relation discovered in this paper can provide insights to the
universal neural architecture search.
Related papers
- Informed deep hierarchical classification: a non-standard analysis inspired approach [0.0]
It consists in a multi-output deep neural network equipped with specific projection operators placed before each output layer.
The design of such an architecture, called lexicographic hybrid deep neural network (LH-DNN), has been possible by combining tools from different and quite distant research fields.
To assess the efficacy of the approach, the resulting network is compared against the B-CNN, a convolutional neural network tailored for hierarchical classification tasks.
arXiv Detail & Related papers (2024-09-25T14:12:50Z) - Towards Optimal Customized Architecture for Heterogeneous Federated
Learning with Contrastive Cloud-Edge Model Decoupling [20.593232086762665]
Federated learning, as a promising distributed learning paradigm, enables collaborative training of a global model across multiple network edge clients without the need for central data collecting.
We propose a novel federated learning framework called FedCMD, a model decoupling tailored to the Cloud-edge supported federated learning.
Our motivation is that, by the deep investigation of the performance of selecting different neural network layers as the personalized head, we found rigidly assigning the last layer as the personalized head in current studies is not always optimal.
arXiv Detail & Related papers (2024-03-04T05:10:28Z) - Multiobjective Evolutionary Pruning of Deep Neural Networks with
Transfer Learning for improving their Performance and Robustness [15.29595828816055]
This work proposes MO-EvoPruneDeepTL, a multi-objective evolutionary pruning algorithm.
We use Transfer Learning to adapt the last layers of Deep Neural Networks, by replacing them with sparse layers evolved by a genetic algorithm.
Experiments show that our proposal achieves promising results in all the objectives, and direct relation are presented.
arXiv Detail & Related papers (2023-02-20T19:33:38Z) - Pruning-as-Search: Efficient Neural Architecture Search via Channel
Pruning and Structural Reparameterization [50.50023451369742]
Pruning-as-Search (PaS) is an end-to-end channel pruning method to search out desired sub-network automatically and efficiently.
Our proposed architecture outperforms prior arts by around $1.0%$ top-1 accuracy on ImageNet-1000 classification task.
arXiv Detail & Related papers (2022-06-02T17:58:54Z) - SuperNet in Neural Architecture Search: A Taxonomic Survey [14.037182039950505]
This survey focuses on the supernet optimization that builds a neural network that assembles all the architectures as its sub models by using weight sharing.
We aim to accomplish that by proposing them as solutions to the common challenges found in the literature: data-side optimization, poor rank correlation alleviation, and transferable NAS for a number of deployment scenarios.
arXiv Detail & Related papers (2022-04-08T08:29:52Z) - NAS-Navigator: Visual Steering for Explainable One-Shot Deep Neural
Network Synthesis [53.106414896248246]
We present a framework that allows analysts to effectively build the solution sub-graph space and guide the network search by injecting their domain knowledge.
Applying this technique in an iterative manner allows analysts to converge to the best performing neural network architecture for a given application.
arXiv Detail & Related papers (2020-09-28T01:48:45Z) - DC-NAS: Divide-and-Conquer Neural Architecture Search [108.57785531758076]
We present a divide-and-conquer (DC) approach to effectively and efficiently search deep neural architectures.
We achieve a $75.1%$ top-1 accuracy on the ImageNet dataset, which is higher than that of state-of-the-art methods using the same search space.
arXiv Detail & Related papers (2020-05-29T09:02:16Z) - A Semi-Supervised Assessor of Neural Architectures [157.76189339451565]
We employ an auto-encoder to discover meaningful representations of neural architectures.
A graph convolutional neural network is introduced to predict the performance of architectures.
arXiv Detail & Related papers (2020-05-14T09:02:33Z) - Fitting the Search Space of Weight-sharing NAS with Graph Convolutional
Networks [100.14670789581811]
We train a graph convolutional network to fit the performance of sampled sub-networks.
With this strategy, we achieve a higher rank correlation coefficient in the selected set of candidates.
arXiv Detail & Related papers (2020-04-17T19:12:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.