SONATA: Self-adaptive Evolutionary Framework for Hardware-aware Neural
Architecture Search
- URL: http://arxiv.org/abs/2402.13204v1
- Date: Tue, 20 Feb 2024 18:15:11 GMT
- Title: SONATA: Self-adaptive Evolutionary Framework for Hardware-aware Neural
Architecture Search
- Authors: Halima Bouzidi, Smail Niar, Hamza Ouarnoughi, El-Ghazali Talbi
- Abstract summary: HW-aware Neural Architecture Search (HW-aware NAS) emerges as an attractive strategy to automate the design of NN.
We propose SONATA, a self-adaptive evolutionary algorithm for HW-aware NAS.
Our method leverages adaptive evolutionary operators guided by the learned importance of NN design parameters.
- Score: 0.7646713951724011
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent advancements in Artificial Intelligence (AI), driven by Neural
Networks (NN), demand innovative neural architecture designs, particularly
within the constrained environments of Internet of Things (IoT) systems, to
balance performance and efficiency. HW-aware Neural Architecture Search
(HW-aware NAS) emerges as an attractive strategy to automate the design of NN
using multi-objective optimization approaches, such as evolutionary algorithms.
However, the intricate relationship between NN design parameters and HW-aware
NAS optimization objectives remains an underexplored research area, overlooking
opportunities to effectively leverage this knowledge to guide the search
process accordingly. Furthermore, the large amount of evaluation data produced
during the search holds untapped potential for refining the optimization
strategy and improving the approximation of the Pareto front. Addressing these
issues, we propose SONATA, a self-adaptive evolutionary algorithm for HW-aware
NAS. Our method leverages adaptive evolutionary operators guided by the learned
importance of NN design parameters. Specifically, through tree-based surrogate
models and a Reinforcement Learning agent, we aspire to gather knowledge on
'How' and 'When' to evolve NN architectures. Comprehensive evaluations across
various NAS search spaces and hardware devices on the ImageNet-1k dataset have
shown the merit of SONATA with up to 0.25% improvement in accuracy and up to
2.42x gains in latency and energy. Our SONATA has seen up to sim$93.6% Pareto
dominance over the native NSGA-II, further stipulating the importance of
self-adaptive evolution operators in HW-aware NAS.
Related papers
- Task-Oriented Real-time Visual Inference for IoVT Systems: A Co-design Framework of Neural Networks and Edge Deployment [61.20689382879937]
Task-oriented edge computing addresses this by shifting data analysis to the edge.
Existing methods struggle to balance high model performance with low resource consumption.
We propose a novel co-design framework to optimize neural network architecture.
arXiv Detail & Related papers (2024-10-29T19:02:54Z) - A Pairwise Comparison Relation-assisted Multi-objective Evolutionary Neural Architecture Search Method with Multi-population Mechanism [58.855741970337675]
Neural architecture search (NAS) enables re-searchers to automatically explore vast search spaces and find efficient neural networks.
NAS suffers from a key bottleneck, i.e., numerous architectures need to be evaluated during the search process.
We propose the SMEM-NAS, a pairwise com-parison relation-assisted multi-objective evolutionary algorithm based on a multi-population mechanism.
arXiv Detail & Related papers (2024-07-22T12:46:22Z) - Evolution and Efficiency in Neural Architecture Search: Bridging the Gap Between Expert Design and Automated Optimization [1.7385545432331702]
The paper provides a comprehensive overview of Neural Architecture Search.
It emphasizes its evolution from manual design to automated, computationally-driven approaches.
It highlights its application across various domains, including medical imaging and natural language processing.
arXiv Detail & Related papers (2024-02-11T18:27:29Z) - DONNAv2 -- Lightweight Neural Architecture Search for Vision tasks [6.628409795264665]
We present the next-generation neural architecture design for computationally efficient neural architecture distillation - DONNAv2.
DONNAv2 reduces the computational cost of DONNA by 10x for the larger datasets.
To improve the quality of NAS search space, DONNAv2 leverages a block knowledge distillation filter to remove blocks with high inference costs.
arXiv Detail & Related papers (2023-09-26T04:48:50Z) - A Hybrid Neural Coding Approach for Pattern Recognition with Spiking
Neural Networks [53.31941519245432]
Brain-inspired spiking neural networks (SNNs) have demonstrated promising capabilities in solving pattern recognition tasks.
These SNNs are grounded on homogeneous neurons that utilize a uniform neural coding for information representation.
In this study, we argue that SNN architectures should be holistically designed to incorporate heterogeneous coding schemes.
arXiv Detail & Related papers (2023-05-26T02:52:12Z) - Enhancing Once-For-All: A Study on Parallel Blocks, Skip Connections and
Early Exits [7.0895962209555465]
Once-For-All (OFA) is an eco-friendly algorithm characterised by the ability to generate easily adaptable models.
OFA is improved from an architectural point of view by including early exits, parallel blocks and dense skip connections.
OFAAv2 improves its accuracy performance on the Tiny ImageNet dataset by up to 12.07% compared to the original version of OFA.
arXiv Detail & Related papers (2023-02-03T17:53:40Z) - HiveNAS: Neural Architecture Search using Artificial Bee Colony
Optimization [0.0]
In this study, we evaluate the viability of Artificial Bee Colony optimization for Neural Architecture Search.
Our proposed framework, HiveNAS, outperforms existing state-of-the-art Swarm Intelligence-based NAS frameworks in a fraction of the time.
arXiv Detail & Related papers (2022-11-18T14:11:47Z) - Neural Architecture Search for Speech Emotion Recognition [72.1966266171951]
We propose to apply neural architecture search (NAS) techniques to automatically configure the SER models.
We show that NAS can improve SER performance (54.89% to 56.28%) while maintaining model parameter sizes.
arXiv Detail & Related papers (2022-03-31T10:16:10Z) - PV-NAS: Practical Neural Architecture Search for Video Recognition [83.77236063613579]
Deep neural networks for video tasks is highly customized and the design of such networks requires domain experts and costly trial and error tests.
Recent advance in network architecture search has boosted the image recognition performance in a large margin.
In this study, we propose a practical solution, namely Practical Video Neural Architecture Search (PV-NAS)
arXiv Detail & Related papers (2020-11-02T08:50:23Z) - MS-RANAS: Multi-Scale Resource-Aware Neural Architecture Search [94.80212602202518]
We propose Multi-Scale Resource-Aware Neural Architecture Search (MS-RANAS)
We employ a one-shot architecture search approach in order to obtain a reduced search cost.
We achieve state-of-the-art results in terms of accuracy-speed trade-off.
arXiv Detail & Related papers (2020-09-29T11:56:01Z) - NASS: Optimizing Secure Inference via Neural Architecture Search [21.72469549507192]
We propose NASS, an integrated framework to search for tailored NN architectures designed specifically for secure inference (SI)
We show that we can achieve the best of both worlds by using NASS, where the prediction accuracy can be improved from 81.6% to 84.6%, while the inference runtime is reduced by 2x and communication bandwidth by 1.9x on the CIFAR-10 dataset.
arXiv Detail & Related papers (2020-01-30T06:37:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.