11 TeraFLOPs per second photonic convolutional accelerator for deep
learning optical neural networks
- URL: http://arxiv.org/abs/2011.07393v1
- Date: Sat, 14 Nov 2020 21:24:01 GMT
- Title: 11 TeraFLOPs per second photonic convolutional accelerator for deep
learning optical neural networks
- Authors: Xingyuan Xu, Mengxi Tan, Bill Corcoran, Jiayang Wu, Andreas Boes,
Thach G. Nguyen, Sai T. Chu, Brent E. Little, Damien G. Hicks, Roberto
Morandotti, Arnan Mitchell, and David J. Moss
- Abstract summary: We demonstrate a universal optical vector convolutional accelerator operating beyond 10 TeraFLOPS (floating point operations per second)
We then use the same hardware to sequentially form a deep optical CNN with ten output neurons, achieving successful recognition of full 10 digits with 900 pixel handwritten digit images with 88% accuracy.
This approach is scalable and trainable to much more complex networks for demanding applications such as unmanned vehicle and real-time video recognition.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Convolutional neural networks (CNNs), inspired by biological visual cortex
systems, are a powerful category of artificial neural networks that can extract
the hierarchical features of raw data to greatly reduce the network parametric
complexity and enhance the predicting accuracy. They are of significant
interest for machine learning tasks such as computer vision, speech
recognition, playing board games and medical diagnosis. Optical neural networks
offer the promise of dramatically accelerating computing speed to overcome the
inherent bandwidth bottleneck of electronics. Here, we demonstrate a universal
optical vector convolutional accelerator operating beyond 10 TeraFLOPS
(floating point operations per second), generating convolutions of images of
250,000 pixels with 8 bit resolution for 10 kernels simultaneously, enough for
facial image recognition. We then use the same hardware to sequentially form a
deep optical CNN with ten output neurons, achieving successful recognition of
full 10 digits with 900 pixel handwritten digit images with 88% accuracy. Our
results are based on simultaneously interleaving temporal, wavelength and
spatial dimensions enabled by an integrated microcomb source. This approach is
scalable and trainable to much more complex networks for demanding applications
such as unmanned vehicle and real-time video recognition.
Related papers
- Transferable polychromatic optical encoder for neural networks [13.311727599288524]
In this paper, we demonstrate an optical encoder that can perform convolution simultaneously in three color channels during the image capture.
Such an optical encoding results in 24,000 times reduction in computational operations, with a state-of-the art classification accuracy (73.2%) in free-space optical system.
arXiv Detail & Related papers (2024-11-05T00:49:47Z) - Optical training of large-scale Transformers and deep neural networks with direct feedback alignment [48.90869997343841]
We experimentally implement a versatile and scalable training algorithm, called direct feedback alignment, on a hybrid electronic-photonic platform.
An optical processing unit performs large-scale random matrix multiplications, which is the central operation of this algorithm, at speeds up to 1500 TeraOps.
We study the compute scaling of our hybrid optical approach, and demonstrate a potential advantage for ultra-deep and wide neural networks.
arXiv Detail & Related papers (2024-09-01T12:48:47Z) - Spatially Varying Nanophotonic Neural Networks [39.1303097259564]
Photonic processors that execute operations using photons instead of electrons promise to enable optical neural networks with ultra-low latency and power consumption.
Existing optical neural networks, limited by the underlying network designs, have achieved image recognition accuracy far below that of state-of-the-art electronic neural networks.
arXiv Detail & Related papers (2023-08-07T08:48:46Z) - EDeNN: Event Decay Neural Networks for low latency vision [26.784944204163363]
We develop a new type of neural network which operates closer to the original event data stream.
We demonstrate state-of-the-art performance in angular velocity regression and competitive optical flow estimation.
arXiv Detail & Related papers (2022-09-09T15:51:39Z) - Variable Bitrate Neural Fields [75.24672452527795]
We present a dictionary method for compressing feature grids, reducing their memory consumption by up to 100x.
We formulate the dictionary optimization as a vector-quantized auto-decoder problem which lets us learn end-to-end discrete neural representations in a space where no direct supervision is available.
arXiv Detail & Related papers (2022-06-15T17:58:34Z) - A quantum algorithm for training wide and deep classical neural networks [72.2614468437919]
We show that conditions amenable to classical trainability via gradient descent coincide with those necessary for efficiently solving quantum linear systems.
We numerically demonstrate that the MNIST image dataset satisfies such conditions.
We provide empirical evidence for $O(log n)$ training of a convolutional neural network with pooling.
arXiv Detail & Related papers (2021-07-19T23:41:03Z) - Quantized Neural Networks via {-1, +1} Encoding Decomposition and
Acceleration [83.84684675841167]
We propose a novel encoding scheme using -1, +1 to decompose quantized neural networks (QNNs) into multi-branch binary networks.
We validate the effectiveness of our method on large-scale image classification, object detection, and semantic segmentation tasks.
arXiv Detail & Related papers (2021-06-18T03:11:15Z) - An optical neural network using less than 1 photon per multiplication [4.003843776219224]
We experimentally demonstrate an optical neural network achieving 99% accuracy on handwritten-digit classification.
This performance was achieved using a custom free-space optical processor.
Our results provide a proof-of-principle for low-optical-power operation.
arXiv Detail & Related papers (2021-04-27T20:43:23Z) - Learning from Event Cameras with Sparse Spiking Convolutional Neural
Networks [0.0]
Convolutional neural networks (CNNs) are now the de facto solution for computer vision problems.
We propose an end-to-end biologically inspired approach using event cameras and spiking neural networks (SNNs)
Our method enables the training of sparse spiking neural networks directly on event data, using the popular deep learning framework PyTorch.
arXiv Detail & Related papers (2021-04-26T13:52:01Z) - Neural Sparse Representation for Image Restoration [116.72107034624344]
Inspired by the robustness and efficiency of sparse coding based image restoration models, we investigate the sparsity of neurons in deep networks.
Our method structurally enforces sparsity constraints upon hidden neurons.
Experiments show that sparse representation is crucial in deep neural networks for multiple image restoration tasks.
arXiv Detail & Related papers (2020-06-08T05:15:17Z) - Computational optimization of convolutional neural networks using
separated filters architecture [69.73393478582027]
We consider a convolutional neural network transformation that reduces computation complexity and thus speedups neural network processing.
Use of convolutional neural networks (CNN) is the standard approach to image recognition despite the fact they can be too computationally demanding.
arXiv Detail & Related papers (2020-02-18T17:42:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.