EDD: Efficient Differentiable DNN Architecture and Implementation
Co-search for Embedded AI Solutions
- URL: http://arxiv.org/abs/2005.02563v1
- Date: Wed, 6 May 2020 02:37:48 GMT
- Title: EDD: Efficient Differentiable DNN Architecture and Implementation
Co-search for Embedded AI Solutions
- Authors: Yuhong Li, Cong Hao, Xiaofan Zhang, Xinheng Liu, Yao Chen, Jinjun
Xiong, Wen-mei Hwu, Deming Chen
- Abstract summary: We propose a fully simultaneous, efficient differentiable DNN architecture and implementation co-search (EDD) methodology.
We formulate the co-search problem by fusing search variables and hardware implementation variables into one solution space, and maximize both algorithm accuracy and hardware implementation quality.
- Score: 40.32848001349242
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: High quality AI solutions require joint optimization of AI algorithms and
their hardware implementations. In this work, we are the first to propose a
fully simultaneous, efficient differentiable DNN architecture and
implementation co-search (EDD) methodology. We formulate the co-search problem
by fusing DNN search variables and hardware implementation variables into one
solution space, and maximize both algorithm accuracy and hardware
implementation quality. The formulation is differentiable with respect to the
fused variables, so that gradient descent algorithm can be applied to greatly
reduce the search time. The formulation is also applicable for various devices
with different objectives. In the experiments, we demonstrate the effectiveness
of our EDD methodology by searching for three representative DNNs, targeting
low-latency GPU implementation and FPGA implementations with both recursive and
pipelined architectures. Each model produced by EDD achieves similar accuracy
as the best existing DNN models searched by neural architecture search (NAS)
methods on ImageNet, but with superior performance obtained within 12 GPU-hour
searches. Our DNN targeting GPU is 1.40x faster than the state-of-the-art
solution reported in Proxyless, and our DNN targeting FPGA delivers 1.45x
higher throughput than the state-of-the-art solution reported in DNNBuilder.
Related papers
- PLiNIO: A User-Friendly Library of Gradient-based Methods for
Complexity-aware DNN Optimization [3.460496851517031]
PLiNIO is an open-source library implementing a comprehensive set of state-of-the-art DNN design automation techniques.
We show that PLiNIO achieves up to 94.34% memory reduction for a 1% accuracy drop compared to a baseline architecture.
arXiv Detail & Related papers (2023-07-18T07:11:14Z) - Flexible Channel Dimensions for Differentiable Architecture Search [50.33956216274694]
We propose a novel differentiable neural architecture search method with an efficient dynamic channel allocation algorithm.
We show that the proposed framework is able to find DNN architectures that are equivalent to previous methods in task accuracy and inference latency.
arXiv Detail & Related papers (2023-06-13T15:21:38Z) - Automatic Mapping of the Best-Suited DNN Pruning Schemes for Real-Time
Mobile Acceleration [71.80326738527734]
We propose a general, fine-grained structured pruning scheme and corresponding compiler optimizations.
We show that our pruning scheme mapping methods, together with the general fine-grained structured pruning scheme, outperform the state-of-the-art DNN optimization framework.
arXiv Detail & Related papers (2021-11-22T23:53:14Z) - HAO: Hardware-aware neural Architecture Optimization for Efficient
Inference [25.265181492143107]
We develop an integer programming algorithm to prune the design space of a neural network search algorithm.
Our algorithm achieves 72.5% top-1 accuracy on ImageNet at framerate 50, which is 60% faster than MnasNet and 135% faster than FBNet with comparable accuracy.
arXiv Detail & Related papers (2021-04-26T17:59:29Z) - OPANAS: One-Shot Path Aggregation Network Architecture Search for Object
Detection [82.04372532783931]
Recently, neural architecture search (NAS) has been exploited to design feature pyramid networks (FPNs)
We propose a novel One-Shot Path Aggregation Network Architecture Search (OPANAS) algorithm, which significantly improves both searching efficiency and detection accuracy.
arXiv Detail & Related papers (2021-03-08T01:48:53Z) - ISTA-NAS: Efficient and Consistent Neural Architecture Search by Sparse
Coding [86.40042104698792]
We formulate neural architecture search as a sparse coding problem.
In experiments, our two-stage method on CIFAR-10 requires only 0.05 GPU-day for search.
Our one-stage method produces state-of-the-art performances on both CIFAR-10 and ImageNet at the cost of only evaluation time.
arXiv Detail & Related papers (2020-10-13T04:34:24Z) - AlphaGAN: Fully Differentiable Architecture Search for Generative
Adversarial Networks [15.740179244963116]
Generative Adversarial Networks (GANs) are formulated as minimax game problems, whereby generators attempt to approach real data distributions by virtue of adversarial learning against discriminators.
In this work, we aim to boost model learning from the perspective of network architectures, by incorporating recent progress on automated architecture search into GANs.
We propose a fully differentiable search framework for generative adversarial networks, dubbed alphaGAN.
arXiv Detail & Related papers (2020-06-16T13:27:30Z) - DDPNAS: Efficient Neural Architecture Search via Dynamic Distribution
Pruning [135.27931587381596]
We propose an efficient and unified NAS framework termed DDPNAS via dynamic distribution pruning.
In particular, we first sample architectures from a joint categorical distribution. Then the search space is dynamically pruned and its distribution is updated every few epochs.
With the proposed efficient network generation method, we directly obtain the optimal neural architectures on given constraints.
arXiv Detail & Related papers (2019-05-28T06:35:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.