NASA: Neural Architecture Search and Acceleration for Hardware Inspired
Hybrid Networks
- URL: http://arxiv.org/abs/2210.13361v1
- Date: Mon, 24 Oct 2022 16:03:42 GMT
- Title: NASA: Neural Architecture Search and Acceleration for Hardware Inspired
Hybrid Networks
- Authors: Huihong Shi, Haoran You, Yang Zhao, Zhongfeng Wang, and Yingyan Lin
- Abstract summary: We propose a Neural Architecture Search and Acceleration framework dubbed NASA.
It enables automated multiplication-reduced DNN development and integrates a dedicated multiplication-reduced accelerator.
Experiments consistently validate the advantages of NASA's algorithm-hardware co-design framework in terms of achievable accuracy and efficiency tradeoffs.
- Score: 24.95135135092478
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multiplication is arguably the most cost-dominant operation in modern deep
neural networks (DNNs), limiting their achievable efficiency and thus more
extensive deployment in resource-constrained applications. To tackle this
limitation, pioneering works have developed handcrafted multiplication-free
DNNs, which require expert knowledge and time-consuming manual iteration,
calling for fast development tools. To this end, we propose a Neural
Architecture Search and Acceleration framework dubbed NASA, which enables
automated multiplication-reduced DNN development and integrates a dedicated
multiplication-reduced accelerator for boosting DNNs' achievable efficiency.
Specifically, NASA adopts neural architecture search (NAS) spaces that augment
the state-of-the-art one with hardware-inspired multiplication-free operators,
such as shift and adder, armed with a novel progressive pretrain strategy (PGP)
together with customized training recipes to automatically search for optimal
multiplication-reduced DNNs; On top of that, NASA further develops a dedicated
accelerator, which advocates a chunk-based template and auto-mapper dedicated
for NASA-NAS resulting DNNs to better leverage their algorithmic properties for
boosting hardware efficiency. Experimental results and ablation studies
consistently validate the advantages of NASA's algorithm-hardware co-design
framework in terms of achievable accuracy and efficiency tradeoffs. Codes are
available at https://github.com/RICE-EIC/NASA.
Related papers
- HASNAS: A Hardware-Aware Spiking Neural Architecture Search Framework for Neuromorphic Compute-in-Memory Systems [6.006032394972252]
Spiking Neural Networks (SNNs) have shown capabilities for solving diverse machine learning tasks with ultra-low-power/energy computation.
We propose HASNAS, a novel hardware-aware spiking neural architecture search framework for neuromorphic CIM systems.
arXiv Detail & Related papers (2024-06-30T09:51:58Z) - Spyx: A Library for Just-In-Time Compiled Optimization of Spiking Neural
Networks [0.08965418284317034]
Spiking Neural Networks (SNNs) offer to enhance energy efficiency through a reduced and low-power hardware footprint.
This paper introduces Spyx, a new and lightweight SNN simulation and optimization library designed in JAX.
arXiv Detail & Related papers (2024-02-29T09:46:44Z) - LitE-SNN: Designing Lightweight and Efficient Spiking Neural Network through Spatial-Temporal Compressive Network Search and Joint Optimization [48.41286573672824]
Spiking Neural Networks (SNNs) mimic the information-processing mechanisms of the human brain and are highly energy-efficient.
We propose a new approach named LitE-SNN that incorporates both spatial and temporal compression into the automated network design process.
arXiv Detail & Related papers (2024-01-26T05:23:11Z) - SpikingJelly: An open-source machine learning infrastructure platform
for spike-based intelligence [51.6943465041708]
Spiking neural networks (SNNs) aim to realize brain-inspired intelligence on neuromorphic chips with high energy efficiency.
We contribute a full-stack toolkit for pre-processing neuromorphic datasets, building deep SNNs, optimizing their parameters, and deploying SNNs on neuromorphic chips.
arXiv Detail & Related papers (2023-10-25T13:15:17Z) - Energy-Efficient Deployment of Machine Learning Workloads on
Neuromorphic Hardware [0.11744028458220425]
Several edge deep learning hardware accelerators have been released that specifically focus on reducing the power and area consumed by deep neural networks (DNNs)
Spiked neural networks (SNNs) which operate on discrete time-series data have been shown to achieve substantial power reductions when deployed on specialized neuromorphic event-based/asynchronous hardware.
In this work, we provide a general guide to converting pre-trained DNNs into SNNs while also presenting techniques to improve the deployment of converted SNNs on neuromorphic hardware.
arXiv Detail & Related papers (2022-10-10T20:27:19Z) - FPGA-optimized Hardware acceleration for Spiking Neural Networks [69.49429223251178]
This work presents the development of a hardware accelerator for an SNN, with off-line training, applied to an image recognition task.
The design targets a Xilinx Artix-7 FPGA, using in total around the 40% of the available hardware resources.
It reduces the classification time by three orders of magnitude, with a small 4.5% impact on the accuracy, if compared to its software, full precision counterpart.
arXiv Detail & Related papers (2022-01-18T13:59:22Z) - DNA: Differentiable Network-Accelerator Co-Search [36.68587348474986]
We propose DNA, a Differentiable Network-Accelerator co-search framework for automatically searching for matched networks and accelerators.
Specifically, DNA integrates two enablers: (1) a generic design space for DNN accelerators and compatible with DNN frameworks such as PyTorch to enable algorithmic exploration.
Experiments and ablation studies show that the matched networks and accelerators generated by DNA consistently outperform state-of-the-art (SOTA) DNNs and accelerators.
arXiv Detail & Related papers (2020-10-28T05:57:16Z) - MS-RANAS: Multi-Scale Resource-Aware Neural Architecture Search [94.80212602202518]
We propose Multi-Scale Resource-Aware Neural Architecture Search (MS-RANAS)
We employ a one-shot architecture search approach in order to obtain a reduced search cost.
We achieve state-of-the-art results in terms of accuracy-speed trade-off.
arXiv Detail & Related papers (2020-09-29T11:56:01Z) - Progressive Tandem Learning for Pattern Recognition with Deep Spiking
Neural Networks [80.15411508088522]
Spiking neural networks (SNNs) have shown advantages over traditional artificial neural networks (ANNs) for low latency and high computational efficiency.
We propose a novel ANN-to-SNN conversion and layer-wise learning framework for rapid and efficient pattern recognition.
arXiv Detail & Related papers (2020-07-02T15:38:44Z) - You Only Spike Once: Improving Energy-Efficient Neuromorphic Inference
to ANN-Level Accuracy [51.861168222799186]
Spiking Neural Networks (SNNs) are a type of neuromorphic, or brain-inspired network.
SNNs are sparse, accessing very few weights, and typically only use addition operations instead of the more power-intensive multiply-and-accumulate operations.
In this work, we aim to overcome the limitations of TTFS-encoded neuromorphic systems.
arXiv Detail & Related papers (2020-06-03T15:55:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.