O-HAS: Optical Hardware Accelerator Search for Boosting Both
Acceleration Performance and Development Speed
- URL: http://arxiv.org/abs/2108.07538v1
- Date: Tue, 17 Aug 2021 09:50:14 GMT
- Title: O-HAS: Optical Hardware Accelerator Search for Boosting Both
Acceleration Performance and Development Speed
- Authors: Mengquan Li, Zhongzhi Yu, Yongan Zhang, Yonggan Fu, Yingyan Lin
- Abstract summary: O-HAS consists of two integrated enablers: (1) an O-Cost Predictor, which can accurately yet efficiently predict an optical accelerator's energy and latency based on the DNN model parameters and the optical accelerator design; and (2) an O-Search Engine, which can automatically explore the large design space of optical DNN accelerators.
Experiments and ablation studies consistently validate the effectiveness of both our O-Cost Predictor and O-Search Engine.
- Score: 13.41883640945134
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The recent breakthroughs and prohibitive complexities of Deep Neural Networks
(DNNs) have excited extensive interest in domain-specific DNN accelerators,
among which optical DNN accelerators are particularly promising thanks to their
unprecedented potential of achieving superior performance-per-watt. However,
the development of optical DNN accelerators is much slower than that of
electrical DNN accelerators. One key challenge is that while many techniques
have been developed to facilitate the development of electrical DNN
accelerators, techniques that support or expedite optical DNN accelerator
design remain much less explored, limiting both the achievable performance and
the innovation development of optical DNN accelerators. To this end, we develop
the first-of-its-kind framework dubbed O-HAS, which for the first time
demonstrates automated Optical Hardware Accelerator Search for boosting both
the acceleration efficiency and development speed of optical DNN accelerators.
Specifically, our O-HAS consists of two integrated enablers: (1) an O-Cost
Predictor, which can accurately yet efficiently predict an optical
accelerator's energy and latency based on the DNN model parameters and the
optical accelerator design; and (2) an O-Search Engine, which can automatically
explore the large design space of optical DNN accelerators and identify the
optimal accelerators (i.e., the micro-architectures and
algorithm-to-accelerator mapping methods) in order to maximize the target
acceleration efficiency. Extensive experiments and ablation studies
consistently validate the effectiveness of both our O-Cost Predictor and
O-Search Engine as well as the excellent efficiency of O-HAS generated optical
accelerators.
Related papers
- DCP: Learning Accelerator Dataflow for Neural Network via Propagation [52.06154296196845]
This work proposes an efficient data-centric approach, named Dataflow Code Propagation (DCP), to automatically find the optimal dataflow for DNN layers in seconds without human effort.
DCP learns a neural predictor to efficiently update the dataflow codes towards the desired gradient directions to minimize various optimization objectives.
For example, without using additional training data, DCP surpasses the GAMMA method that performs a full search using thousands of samples.
arXiv Detail & Related papers (2024-10-09T05:16:44Z) - EPIM: Efficient Processing-In-Memory Accelerators based on Epitome [78.79382890789607]
We introduce the Epitome, a lightweight neural operator offering convolution-like functionality.
On the software side, we evaluate epitomes' latency and energy on PIM accelerators.
We introduce a PIM-aware layer-wise design method to enhance their hardware efficiency.
arXiv Detail & Related papers (2023-11-12T17:56:39Z) - SpikingJelly: An open-source machine learning infrastructure platform
for spike-based intelligence [51.6943465041708]
Spiking neural networks (SNNs) aim to realize brain-inspired intelligence on neuromorphic chips with high energy efficiency.
We contribute a full-stack toolkit for pre-processing neuromorphic datasets, building deep SNNs, optimizing their parameters, and deploying SNNs on neuromorphic chips.
arXiv Detail & Related papers (2023-10-25T13:15:17Z) - NASA: Neural Architecture Search and Acceleration for Hardware Inspired
Hybrid Networks [24.95135135092478]
We propose a Neural Architecture Search and Acceleration framework dubbed NASA.
It enables automated multiplication-reduced DNN development and integrates a dedicated multiplication-reduced accelerator.
Experiments consistently validate the advantages of NASA's algorithm-hardware co-design framework in terms of achievable accuracy and efficiency tradeoffs.
arXiv Detail & Related papers (2022-10-24T16:03:42Z) - FPGA-optimized Hardware acceleration for Spiking Neural Networks [69.49429223251178]
This work presents the development of a hardware accelerator for an SNN, with off-line training, applied to an image recognition task.
The design targets a Xilinx Artix-7 FPGA, using in total around the 40% of the available hardware resources.
It reduces the classification time by three orders of magnitude, with a small 4.5% impact on the accuracy, if compared to its software, full precision counterpart.
arXiv Detail & Related papers (2022-01-18T13:59:22Z) - 2-in-1 Accelerator: Enabling Random Precision Switch for Winning Both
Adversarial Robustness and Efficiency [26.920864182619844]
We propose a 2-in-1 Accelerator aiming at winning both the adversarial robustness and efficiency of DNN accelerators.
Specifically, we first propose a Random Precision Switch (RPS) algorithm that can effectively defend DNNs against adversarial attacks.
Furthermore, we propose a new precision-scalable accelerator featuring (1) a new precision-scalable unit architecture.
arXiv Detail & Related papers (2021-09-11T08:51:01Z) - ROBIN: A Robust Optical Binary Neural Network Accelerator [3.8137985834223507]
Domain specific neural network accelerators have garnered attention because of their improved energy efficiency and inference performance.
We present a novel optical-domain BNN accelerator, named ROBIN, which intelligently integrates heterogeneous microring resonator optical devices.
Our analysis shows that ROBIN can outperform the best-known optical BNN accelerators and also many electronic accelerators.
arXiv Detail & Related papers (2021-07-12T16:00:32Z) - InstantNet: Automated Generation and Deployment of Instantaneously
Switchable-Precision Networks [65.78061366594106]
We propose InstantNet to automatically generate and deploy instantaneously switchable-precision networks which operate at variable bit-widths.
In experiments, the proposed InstantNet consistently outperforms state-of-the-art designs.
arXiv Detail & Related papers (2021-04-22T04:07:43Z) - DNA: Differentiable Network-Accelerator Co-Search [36.68587348474986]
We propose DNA, a Differentiable Network-Accelerator co-search framework for automatically searching for matched networks and accelerators.
Specifically, DNA integrates two enablers: (1) a generic design space for DNN accelerators and compatible with DNN frameworks such as PyTorch to enable algorithmic exploration.
Experiments and ablation studies show that the matched networks and accelerators generated by DNA consistently outperform state-of-the-art (SOTA) DNNs and accelerators.
arXiv Detail & Related papers (2020-10-28T05:57:16Z) - DNN-Chip Predictor: An Analytical Performance Predictor for DNN
Accelerators with Various Dataflows and Hardware Architectures [30.689015188050405]
The recent breakthroughs in deep neural networks (DNNs) have spurred a tremendously increased demand for DNN accelerators.
DNN-Chip Predictor is an analytical performance predictor which can accurately predict DNN accelerators' energy, throughput, and latency prior to their actual implementation.
arXiv Detail & Related papers (2020-02-26T02:59:18Z) - Temporal Pulses Driven Spiking Neural Network for Fast Object
Recognition in Autonomous Driving [65.36115045035903]
We propose an approach to address the object recognition problem directly with raw temporal pulses utilizing the spiking neural network (SNN)
Being evaluated on various datasets, our proposed method has shown comparable performance as the state-of-the-art methods, while achieving remarkable time efficiency.
arXiv Detail & Related papers (2020-01-24T22:58:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.