Batch Processing and Data Streaming Fourier-based Convolutional Neural
Network Accelerator
- URL: http://arxiv.org/abs/2112.12297v1
- Date: Thu, 23 Dec 2021 01:06:17 GMT
- Title: Batch Processing and Data Streaming Fourier-based Convolutional Neural
Network Accelerator
- Authors: Zibo Hu, Shurui Li, Russell L.T. Schwartz, Maria Solyanik-Gorgone,
Mario Miscuglio, Puneet Gupta, Volker J. Sorger
- Abstract summary: Decision-making by artificial neural networks with minimal latency is paramount for numerous applications such as navigation, tracking, and real-time machine action systems.
This requires the machine learning hardware to handle multidimensional data with a high throughput.
We demonstrate a non-van Neuman-based machine learning acceleration with a Fourier Convolutional Neural Network (FCNN) accelerator.
- Score: 4.7257913147626995
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Decision-making by artificial neural networks with minimal latency is
paramount for numerous applications such as navigation, tracking, and real-time
machine action systems. This requires the machine learning hardware to handle
multidimensional data with a high throughput. Processing convolution operations
being the major computational tool for data classification tasks,
unfortunately, follows a challenging run-time complexity scaling law. However,
implementing the convolution theorem homomorphically in a Fourier-optic
display-light-processor enables a non-iterative O(1) runtime complexity for
data inputs beyond 1,000 x 1,000 large matrices. Following this approach, here
we demonstrate data streaming multi-kernel image batch-processing with a
Fourier Convolutional Neural Network (FCNN) accelerator. We show image batch
processing of large-scale matrices as passive 2-million dot-product
multiplications performed by digital light-processing modules in the Fourier
domain. In addition, we parallelize this optical FCNN system further by
utilizing multiple spatio-parallel diffraction orders, thus achieving a
98-times throughput improvement over state-of-art FCNN accelerators. The
comprehensive discussion of the practical challenges related to working on the
edge of the system's capabilities highlights issues of crosstalk in the Fourier
domain and resolution scaling laws. Accelerating convolutions by utilizing the
massive parallelism in display technology brings forth a non-van Neuman-based
machine learning acceleration.
Related papers
- Neuromorphic Wireless Split Computing with Multi-Level Spikes [69.73249913506042]
In neuromorphic computing, spiking neural networks (SNNs) perform inference tasks, offering significant efficiency gains for workloads involving sequential data.
Recent advances in hardware and software have demonstrated that embedding a few bits of payload in each spike exchanged between the spiking neurons can further enhance inference accuracy.
This paper investigates a wireless neuromorphic split computing architecture employing multi-level SNNs.
arXiv Detail & Related papers (2024-11-07T14:08:35Z) - Adaptive Frequency Filters As Efficient Global Token Mixers [100.27957692579892]
We show that adaptive frequency filters can serve as efficient global token mixers.
We take AFF token mixers as primary neural operators to build a lightweight neural network, dubbed AFFNet.
arXiv Detail & Related papers (2023-07-26T07:42:28Z) - Scaling Up Computer Vision Neural Networks Using Fast Fourier Transform [0.0]
Deep Learning-based Computer Vision field has recently been trying to explore larger kernels for convolution to effectively scale up Convolutional Neural Networks.
New paradigm of models such as Vision Transformers find it difficult to scale up to larger higher resolution images due to their quadratic complexity in terms of input sequence.
arXiv Detail & Related papers (2023-02-02T19:19:10Z) - PARTIME: Scalable and Parallel Processing Over Time with Deep Neural
Networks [68.96484488899901]
We present PARTIME, a library designed to speed up neural networks whenever data is continuously streamed over time.
PARTIME starts processing each data sample at the time in which it becomes available from the stream.
Experiments are performed in order to empirically compare PARTIME with classic non-parallel neural computations in online learning.
arXiv Detail & Related papers (2022-10-17T14:49:14Z) - RF-Photonic Deep Learning Processor with Shannon-Limited Data Movement [0.0]
Optical neural networks (ONNs) are promising accelerators with ultra-low latency and energy consumption.
We introduce our multiplicative analog frequency transform ONN (MAFT-ONN) that encodes the data in the frequency domain.
We experimentally demonstrate the first hardware accelerator that computes fully-analog deep learning on raw RF signals.
arXiv Detail & Related papers (2022-07-08T16:37:13Z) - Time-coded Spiking Fourier Transform in Neuromorphic Hardware [4.432142139656578]
In this work, we propose a time-based spiking neural network that is mathematically equivalent tothe Fourier transform.
We implemented the network in the neuromorphic chip Loihi and conductedexperiments on five different real scenarios with an automotive frequency modulated continuouswave radar.
arXiv Detail & Related papers (2022-02-25T12:15:46Z) - Adaptive Fourier Neural Operators: Efficient Token Mixers for
Transformers [55.90468016961356]
We propose an efficient token mixer that learns to mix in the Fourier domain.
AFNO is based on a principled foundation of operator learning.
It can handle a sequence size of 65k and outperforms other efficient self-attention mechanisms.
arXiv Detail & Related papers (2021-11-24T05:44:31Z) - Adaptive Machine Learning for Time-Varying Systems: Low Dimensional
Latent Space Tuning [91.3755431537592]
We present a recently developed method of adaptive machine learning for time-varying systems.
Our approach is to map very high (N>100k) dimensional inputs into the low dimensional (N2) latent space at the output of the encoder section of an encoder-decoder CNN.
This method allows us to learn correlations within and to track their evolution in real time based on feedback without interrupts.
arXiv Detail & Related papers (2021-07-13T16:05:28Z) - Learning Frequency-aware Dynamic Network for Efficient Super-Resolution [56.98668484450857]
This paper explores a novel frequency-aware dynamic network for dividing the input into multiple parts according to its coefficients in the discrete cosine transform (DCT) domain.
In practice, the high-frequency part will be processed using expensive operations and the lower-frequency part is assigned with cheap operations to relieve the computation burden.
Experiments conducted on benchmark SISR models and datasets show that the frequency-aware dynamic network can be employed for various SISR neural architectures.
arXiv Detail & Related papers (2021-03-15T12:54:26Z) - Knowledge Distillation Circumvents Nonlinearity for Optical
Convolutional Neural Networks [4.683612295430957]
We propose a Spectral CNN Linear Counterpart (SCLC) network architecture and develop a Knowledge Distillation (KD) approach to circumvent the need for a nonlinearity.
We show that the KD approach can achieve performance that easily surpasses the standard linear version of a CNN and could approach the performance of the nonlinear network.
arXiv Detail & Related papers (2021-02-26T06:35:34Z) - Efficient Algorithms for Device Placement of DNN Graph Operators [12.871398348743591]
Modern machine learning workloads use large models, with complex structures, that are very expensive to execute.
The devices that execute complex models are becoming increasingly heterogeneous as we see a flourishing of domain-specific accelerators being offered as hardware accelerators in addition to CPUs.
Recent work has shown that significant gains can be obtained with model parallelism, i.e., partitioning a neural network's computational graph onto multiple devices.
In this paper, we identify and isolate the structured optimization problem at the core of device placement of DNN operators, for both inference and training, especially in modern pipelined settings.
arXiv Detail & Related papers (2020-06-29T22:45:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.