NAS-LID: Efficient Neural Architecture Search with Local Intrinsic
Dimension
- URL: http://arxiv.org/abs/2211.12759v2
- Date: Thu, 24 Nov 2022 12:49:54 GMT
- Title: NAS-LID: Efficient Neural Architecture Search with Local Intrinsic
Dimension
- Authors: Xin He, Jiangchao Yao, Yuxin Wang, Zhenheng Tang, Ka Chu Cheung, Simon
See, Bo Han, and Xiaowen Chu
- Abstract summary: One-shot architecture search (NAS) substantially improves the search efficiency by training one supernet to estimate every possible child architecture.
Experiments on NASBench-201 indicate that NAS-LID achieves superior performance with better efficiency.
- Score: 37.04463309816036
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: One-shot neural architecture search (NAS) substantially improves the search
efficiency by training one supernet to estimate the performance of every
possible child architecture (i.e., subnet). However, the inconsistency of
characteristics among subnets incurs serious interference in the optimization,
resulting in poor performance ranking correlation of subnets. Subsequent
explorations decompose supernet weights via a particular criterion, e.g.,
gradient matching, to reduce the interference; yet they suffer from huge
computational cost and low space separability. In this work, we propose a
lightweight and effective local intrinsic dimension (LID)-based method NAS-LID.
NAS-LID evaluates the geometrical properties of architectures by calculating
the low-cost LID features layer-by-layer, and the similarity characterized by
LID enjoys better separability compared with gradients, which thus effectively
reduces the interference among subnets. Extensive experiments on NASBench-201
indicate that NAS-LID achieves superior performance with better efficiency.
Specifically, compared to the gradient-driven method, NAS-LID can save up to
86% of GPU memory overhead when searching on NASBench-201. We also demonstrate
the effectiveness of NAS-LID on ProxylessNAS and OFA spaces. Source code:
https://github.com/marsggbo/NAS-LID.
Related papers
- Efficient Few-Shot Neural Architecture Search by Counting the Number of Nonlinear Functions [29.76210308781724]
We introduce a novel few-shot NAS method that exploits the number of nonlinear functions to split the search space.
Our method is efficient, since it does not require comparing gradients of a supernet to split the space.
In addition, we have found that dividing the space allows us to reduce the channel dimensions required for each supernet.
arXiv Detail & Related papers (2024-12-19T09:31:53Z) - HEP-NAS: Towards Efficient Few-shot Neural Architecture Search via Hierarchical Edge Partitioning [8.484729345263153]
One-shot methods have advanced the field of neural architecture search (NAS) by adopting weight-sharing strategy to reduce search costs.
Few-shot methods divide the entire supernet into individual sub-supernets by splitting edge by edge to alleviate this issue.
We introduce HEP-NAS, a hierarchy-wise partition algorithm designed to further enhance accuracy.
arXiv Detail & Related papers (2024-12-14T07:42:56Z) - Delta-NAS: Difference of Architecture Encoding for Predictor-based Evolutionary Neural Architecture Search [5.1331676121360985]
We craft an algorithm with the capability to perform fine-grain NAS at a low cost.
We propose projecting the problem to a lower dimensional space through predicting the difference in accuracy of a pair of similar networks.
arXiv Detail & Related papers (2024-11-21T02:43:32Z) - Are Neural Architecture Search Benchmarks Well Designed? A Deeper Look
Into Operation Importance [5.065947993017157]
We conduct an empirical analysis of the widely used NAS-Bench-101, NAS-Bench-201 and TransNAS-Bench-101 benchmarks.
We found that only a subset of the operation pool is required to generate architectures close to the upper-bound of the performance range.
We consistently found convolution layers to have the highest impact on the architecture's performance.
arXiv Detail & Related papers (2023-03-29T18:03:28Z) - Generalization Properties of NAS under Activation and Skip Connection
Search [66.8386847112332]
We study the generalization properties of Neural Architecture Search (NAS) under a unifying framework.
We derive the lower (and upper) bounds of the minimum eigenvalue of the Neural Tangent Kernel (NTK) under the (in)finite-width regime.
We show how the derived results can guide NAS to select the top-performing architectures, even in the case without training.
arXiv Detail & Related papers (2022-09-15T12:11:41Z) - Generalizing Few-Shot NAS with Gradient Matching [165.5690495295074]
One-Shot methods train one supernet to approximate the performance of every architecture in the search space via weight-sharing.
Few-Shot NAS reduces the level of weight-sharing by splitting the One-Shot supernet into multiple separated sub-supernets.
It significantly outperforms its Few-Shot counterparts while surpassing previous comparable methods in terms of the accuracy of derived architectures.
arXiv Detail & Related papers (2022-03-29T03:06:16Z) - NAS-FCOS: Efficient Search for Object Detection Architectures [113.47766862146389]
We propose an efficient method to obtain better object detectors by searching for the feature pyramid network (FPN) and the prediction head of a simple anchor-free object detector.
With carefully designed search space, search algorithms, and strategies for evaluating network quality, we are able to find top-performing detection architectures within 4 days using 8 V100 GPUs.
arXiv Detail & Related papers (2021-10-24T12:20:04Z) - iDARTS: Differentiable Architecture Search with Stochastic Implicit
Gradients [75.41173109807735]
Differentiable ARchiTecture Search (DARTS) has recently become the mainstream of neural architecture search (NAS)
We tackle the hypergradient computation in DARTS based on the implicit function theorem.
We show that the architecture optimisation with the proposed method, named iDARTS, is expected to converge to a stationary point.
arXiv Detail & Related papers (2021-06-21T00:44:11Z) - AdvantageNAS: Efficient Neural Architecture Search with Credit
Assignment [23.988393741948485]
We propose a novel search strategy for one-shot and sparse propagation NAS, namely AdvantageNAS.
AdvantageNAS is a gradient-based approach that improves the search efficiency by introducing credit assignment in gradient estimation for architecture updates.
Experiments on the NAS-Bench-201 and PTB dataset show that AdvantageNAS discovers an architecture with higher performance under a limited time budget.
arXiv Detail & Related papers (2020-12-11T05:45:03Z) - Binarized Neural Architecture Search for Efficient Object Recognition [120.23378346337311]
Binarized neural architecture search (BNAS) produces extremely compressed models to reduce huge computational cost on embedded devices for edge computing.
An accuracy of $96.53%$ vs. $97.22%$ is achieved on the CIFAR-10 dataset, but with a significantly compressed model, and a $40%$ faster search than the state-of-the-art PC-DARTS.
arXiv Detail & Related papers (2020-09-08T15:51:23Z) - DDPNAS: Efficient Neural Architecture Search via Dynamic Distribution
Pruning [135.27931587381596]
We propose an efficient and unified NAS framework termed DDPNAS via dynamic distribution pruning.
In particular, we first sample architectures from a joint categorical distribution. Then the search space is dynamically pruned and its distribution is updated every few epochs.
With the proposed efficient network generation method, we directly obtain the optimal neural architectures on given constraints.
arXiv Detail & Related papers (2019-05-28T06:35:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.