Grassroots Operator Search for Model Edge Adaptation
- URL: http://arxiv.org/abs/2309.11246v1
- Date: Wed, 20 Sep 2023 12:15:58 GMT
- Title: Grassroots Operator Search for Model Edge Adaptation
- Authors: Hadjer Benmeziane, Kaoutar El Maghraoui, Hamza Ouarnoughi, Smail Niar
- Abstract summary: Hardware-aware Neural Architecture (HW-NAS) is increasingly being used to design efficient deep learning architectures.
We present a Grassroots Operator Search (GOS) methodology to search for efficient operator replacement.
Our method consistently outperforms the original models on two edge devices, with a minimum of 2.2x speedup while maintaining high accuracy.
- Score: 2.1756721838833797
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Hardware-aware Neural Architecture Search (HW-NAS) is increasingly being used
to design efficient deep learning architectures. An efficient and flexible
search space is crucial to the success of HW-NAS. Current approaches focus on
designing a macro-architecture and searching for the architecture's
hyperparameters based on a set of possible values. This approach is biased by
the expertise of deep learning (DL) engineers and standard modeling approaches.
In this paper, we present a Grassroots Operator Search (GOS) methodology. Our
HW-NAS adapts a given model for edge devices by searching for efficient
operator replacement. We express each operator as a set of mathematical
instructions that capture its behavior. The mathematical instructions are then
used as the basis for searching and selecting efficient replacement operators
that maintain the accuracy of the original model while reducing computational
complexity. Our approach is grassroots since it relies on the mathematical
foundations to construct new and efficient operators for DL architectures. We
demonstrate on various DL models, that our method consistently outperforms the
original models on two edge devices, namely Redmi Note 7S and Raspberry Pi3,
with a minimum of 2.2x speedup while maintaining high accuracy. Additionally,
we showcase a use case of our GOS approach in pulse rate estimation on
wristband devices, where we achieve state-of-the-art performance, while
maintaining reduced computational complexity, demonstrating the effectiveness
of our approach in practical applications.
Related papers
- Lightweight Neural Architecture Search for Temporal Convolutional
Networks at the Edge [21.72253397805102]
This work focuses in particular on Temporal Convolutional Networks (TCNs), a convolutional model for time-series processing.
We propose the first NAS tool that explicitly targets the optimization of the most peculiar architectural parameters of TCNs.
We test the proposed NAS on four real-world, edge-relevant tasks, involving audio and bio-signals.
arXiv Detail & Related papers (2023-01-24T19:47:40Z) - POPNASv3: a Pareto-Optimal Neural Architecture Search Solution for Image
and Time Series Classification [8.190723030003804]
This article presents the third version of a sequential model-based NAS algorithm targeting different hardware environments and multiple classification tasks.
Our method is able to find competitive architectures within large search spaces, while keeping a flexible structure and data processing pipeline to adapt to different tasks.
The experiments performed on images and time series classification datasets provide evidence that POPNASv3 can explore a large set of assorted operators and converge to optimal architectures suited for the type of data provided under different scenarios.
arXiv Detail & Related papers (2022-12-13T17:14:14Z) - POPNASv2: An Efficient Multi-Objective Neural Architecture Search
Technique [7.497722345725035]
This paper proposes a new version of the Pareto-optimal Progressive Neural Architecture Search, called POPNASv2.
Our approach enhances its first version and improves its performance.
Our efforts allow POPNASv2 to achieve PNAS-like performance with an average 4x factor search time speed-up.
arXiv Detail & Related papers (2022-10-06T14:51:54Z) - FlowNAS: Neural Architecture Search for Optical Flow Estimation [65.44079917247369]
We propose a neural architecture search method named FlowNAS to automatically find the better encoder architecture for flow estimation task.
Experimental results show that the discovered architecture with the weights inherited from the super-network achieves 4.67% F1-all error on KITTI.
arXiv Detail & Related papers (2022-07-04T09:05:25Z) - Shapley-NAS: Discovering Operation Contribution for Neural Architecture
Search [96.20505710087392]
We propose a Shapley value based method to evaluate operation contribution (Shapley-NAS) for neural architecture search.
We show that our method outperforms the state-of-the-art methods by a considerable margin with light search cost.
arXiv Detail & Related papers (2022-06-20T14:41:49Z) - ZARTS: On Zero-order Optimization for Neural Architecture Search [94.41017048659664]
Differentiable architecture search (DARTS) has been a popular one-shot paradigm for NAS due to its high efficiency.
This work turns to zero-order optimization and proposes a novel NAS scheme, called ZARTS, to search without enforcing the above approximation.
In particular, results on 12 benchmarks verify the outstanding robustness of ZARTS, where the performance of DARTS collapses due to its known instability issue.
arXiv Detail & Related papers (2021-10-10T09:35:15Z) - Conceptual Expansion Neural Architecture Search (CENAS) [1.3464152928754485]
We present an approach called Conceptual Expansion Neural Architecture Search (CENAS)
It combines a sample-efficient, computational creativity-inspired transfer learning approach with neural architecture search.
It finds models faster than naive architecture search via transferring existing weights to approximate the parameters of the new model.
arXiv Detail & Related papers (2021-10-07T02:29:26Z) - AutoBERT-Zero: Evolving BERT Backbone from Scratch [94.89102524181986]
We propose an Operation-Priority Neural Architecture Search (OP-NAS) algorithm to automatically search for promising hybrid backbone architectures.
We optimize both the search algorithm and evaluation of candidate models to boost the efficiency of our proposed OP-NAS.
Experiments show that the searched architecture (named AutoBERT-Zero) significantly outperforms BERT and its variants of different model capacities in various downstream tasks.
arXiv Detail & Related papers (2021-07-15T16:46:01Z) - iDARTS: Differentiable Architecture Search with Stochastic Implicit
Gradients [75.41173109807735]
Differentiable ARchiTecture Search (DARTS) has recently become the mainstream of neural architecture search (NAS)
We tackle the hypergradient computation in DARTS based on the implicit function theorem.
We show that the architecture optimisation with the proposed method, named iDARTS, is expected to converge to a stationary point.
arXiv Detail & Related papers (2021-06-21T00:44:11Z) - Multi-objective Neural Architecture Search with Almost No Training [9.93048700248444]
We propose an effective alternative, dubbed Random-Weight Evaluation (RWE), to rapidly estimate the performance of network architectures.
RWE reduces the computational cost of evaluating an architecture from hours to seconds.
When integrated within an evolutionary multi-objective algorithm, RWE obtains a set of efficient architectures with state-of-the-art performance on CIFAR-10 with less than two hours' searching on a single GPU card.
arXiv Detail & Related papers (2020-11-27T07:39:17Z) - DDPNAS: Efficient Neural Architecture Search via Dynamic Distribution
Pruning [135.27931587381596]
We propose an efficient and unified NAS framework termed DDPNAS via dynamic distribution pruning.
In particular, we first sample architectures from a joint categorical distribution. Then the search space is dynamically pruned and its distribution is updated every few epochs.
With the proposed efficient network generation method, we directly obtain the optimal neural architectures on given constraints.
arXiv Detail & Related papers (2019-05-28T06:35:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.