Survey of Machine Learning Accelerators
- URL: http://arxiv.org/abs/2009.00993v1
- Date: Tue, 1 Sep 2020 01:28:59 GMT
- Title: Survey of Machine Learning Accelerators
- Authors: Albert Reuther, Peter Michaleas, Michael Jones, Vijay Gadepally,
Siddharth Samsi and Jeremy Kepner
- Abstract summary: This paper updates the survey of of AI accelerators and processors from last year's IEEE-HPEC paper.
This paper collects and summarizes the current accelerators that have been publicly announced with performance and power consumption numbers.
- Score: 15.163544680926474
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: New machine learning accelerators are being announced and released each month
for a variety of applications from speech recognition, video object detection,
assisted driving, and many data center applications. This paper updates the
survey of of AI accelerators and processors from last year's IEEE-HPEC paper.
This paper collects and summarizes the current accelerators that have been
publicly announced with performance and power consumption numbers. The
performance and power values are plotted on a scatter graph and a number of
dimensions and observations from the trends on this plot are discussed and
analyzed. For instance, there are interesting trends in the plot regarding
power consumption, numerical precision, and inference versus training. This
year, there are many more announced accelerators that are implemented with many
more architectures and technologies from vector engines, dataflow engines,
neuromorphic designs, flash-based analog memory processing, and photonic-based
processing.
Related papers
- Using the Abstract Computer Architecture Description Language to Model
AI Hardware Accelerators [77.89070422157178]
Manufacturers of AI-integrated products face a critical challenge: selecting an accelerator that aligns with their product's performance requirements.
The Abstract Computer Architecture Description Language (ACADL) is a concise formalization of computer architecture block diagrams.
In this paper, we demonstrate how to use the ACADL to model AI hardware accelerators, use their ACADL description to map DNNs onto them, and explain the timing simulation semantics to gather performance results.
arXiv Detail & Related papers (2024-01-30T19:27:16Z) - Lincoln AI Computing Survey (LAICS) Update [8.790207519640472]
This paper is an update of the survey of AI accelerators and processors from past four years.
It collects and summarizes the current commercial accelerators that have been publicly announced.
Market segments are highlighted on the scatter plot, and zoomed plots of each segment are also included.
arXiv Detail & Related papers (2023-10-13T14:36:26Z) - A Survey on Deep Learning Hardware Accelerators for Heterogeneous HPC Platforms [9.036774656254375]
This survey summarizes and classifies the most recent advances in designing deep learning accelerators.
It highlights the most advanced approaches to support deep learning accelerations including not only GPU and TPU-based accelerators but also design-specific hardware accelerators.
The survey also describes accelerators based on emerging memory technologies and computing paradigms.
arXiv Detail & Related papers (2023-06-27T15:24:24Z) - Benchmarking GPU and TPU Performance with Graph Neural Networks [0.0]
This work analyzes and compares the GPU and TPU performance training a Graph Neural Network (GNN) developed to solve a real-life pattern recognition problem.
Characterizing the new class of models acting on sparse data may prove helpful in optimizing the design of deep learning libraries and future AI accelerators.
arXiv Detail & Related papers (2022-10-21T21:03:40Z) - Advancing Reacting Flow Simulations with Data-Driven Models [50.9598607067535]
Key to effective use of machine learning tools in multi-physics problems is to couple them to physical and computer models.
The present chapter reviews some of the open opportunities for the application of data-driven reduced-order modeling of combustion systems.
arXiv Detail & Related papers (2022-09-05T16:48:34Z) - Automated Graph Machine Learning: Approaches, Libraries, Benchmarks and Directions [58.220137936626315]
This paper extensively discusses automated graph machine learning approaches.
We introduce AutoGL, our dedicated and the world's first open-source library for automated graph machine learning.
Also, we describe a tailored benchmark that supports unified, reproducible, and efficient evaluations.
arXiv Detail & Related papers (2022-01-04T18:31:31Z) - SOLIS -- The MLOps journey from data acquisition to actionable insights [62.997667081978825]
In this paper we present a unified deployment pipeline and freedom-to-operate approach that supports all requirements while using basic cross-platform tensor framework and script language engines.
This approach however does not supply the needed procedures and pipelines for the actual deployment of machine learning capabilities in real production grade systems.
arXiv Detail & Related papers (2021-12-22T14:45:37Z) - AI Accelerator Survey and Trends [4.722078109242797]
This paper updates the survey of AI accelerators and processors from past two years.
This paper collects and summarizes the current commercial accelerators that have been publicly announced with peak performance and power consumption numbers.
arXiv Detail & Related papers (2021-09-18T15:57:47Z) - Resistive Neural Hardware Accelerators [0.46198289193451136]
ReRAM-based in-memory computing has great potential in the implementation of area and power efficient inference.
The shift towards ReRAM-based in-memory computing has great potential in the implementation of area and power efficient inference.
In this survey, we review the state-of-the-art ReRAM-based Deep Neural Networks (DNNs) many-core accelerators.
arXiv Detail & Related papers (2021-09-08T21:11:48Z) - Automated Machine Learning Techniques for Data Streams [91.3755431537592]
This paper surveys the state-of-the-art open-source AutoML tools, applies them to data collected from streams, and measures how their performance changes over time.
The results show that off-the-shelf AutoML tools can provide satisfactory results but in the presence of concept drift, detection or adaptation techniques have to be applied to maintain the predictive accuracy over time.
arXiv Detail & Related papers (2021-06-14T11:42:46Z) - Knowledge Distillation: A Survey [87.51063304509067]
Deep neural networks have been successful in both industry and academia, especially for computer vision tasks.
It is a challenge to deploy these cumbersome deep models on devices with limited resources.
Knowledge distillation effectively learns a small student model from a large teacher model.
arXiv Detail & Related papers (2020-06-09T21:47:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.