An optical neural network using less than 1 photon per multiplication
- URL: http://arxiv.org/abs/2104.13467v1
- Date: Tue, 27 Apr 2021 20:43:23 GMT
- Title: An optical neural network using less than 1 photon per multiplication
- Authors: Tianyu Wang, Shi-Yuan Ma, Logan G. Wright, Tatsuhiro Onodera, Brian
Richard and Peter L. McMahon
- Abstract summary: We experimentally demonstrate an optical neural network achieving 99% accuracy on handwritten-digit classification.
This performance was achieved using a custom free-space optical processor.
Our results provide a proof-of-principle for low-optical-power operation.
- Score: 4.003843776219224
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning has rapidly become a widespread tool in both scientific and
commercial endeavors. Milestones of deep learning exceeding human performance
have been achieved for a growing number of tasks over the past several years,
across areas as diverse as game-playing, natural-language translation, and
medical-image analysis. However, continued progress is increasingly hampered by
the high energy costs associated with training and running deep neural networks
on electronic processors. Optical neural networks have attracted attention as
an alternative physical platform for deep learning, as it has been
theoretically predicted that they can fundamentally achieve higher energy
efficiency than neural networks deployed on conventional digital computers.
Here, we experimentally demonstrate an optical neural network achieving 99%
accuracy on handwritten-digit classification using ~3.2 detected photons per
weight multiplication and ~90% accuracy using ~0.64 photons (~$2.4 \times
10^{-19}$ J of optical energy) per weight multiplication. This performance was
achieved using a custom free-space optical processor that executes
matrix-vector multiplications in a massively parallel fashion, with up to ~0.5
million scalar (weight) multiplications performed at the same time. Using
commercially available optical components and standard neural-network training
methods, we demonstrated that optical neural networks can operate near the
standard quantum limit with extremely low optical powers and still achieve high
accuracy. Our results provide a proof-of-principle for low-optical-power
operation, and with careful system design including the surrounding electronics
used for data storage and control, open up a path to realizing optical
processors that require only $10^{-16}$ J total energy per scalar
multiplication -- which is orders of magnitude more efficient than current
digital processors.
Related papers
- Optical training of large-scale Transformers and deep neural networks with direct feedback alignment [48.90869997343841]
We experimentally implement a versatile and scalable training algorithm, called direct feedback alignment, on a hybrid electronic-photonic platform.
An optical processing unit performs large-scale random matrix multiplications, which is the central operation of this algorithm, at speeds up to 1500 TeraOps.
We study the compute scaling of our hybrid optical approach, and demonstrate a potential advantage for ultra-deep and wide neural networks.
arXiv Detail & Related papers (2024-09-01T12:48:47Z) - Expressivity of Neural Networks with Random Weights and Learned Biases [44.02417750529102]
Recent work has pushed the bounds of universal approximation by showing that arbitrary functions can similarly be learned by tuning smaller subsets of parameters.
We provide theoretical and numerical evidence demonstrating that feedforward neural networks with fixed random weights can be trained to perform multiple tasks by learning biases only.
Our results are relevant to neuroscience, where they demonstrate the potential for behaviourally relevant changes in dynamics without modifying synaptic weights.
arXiv Detail & Related papers (2024-07-01T04:25:49Z) - Genetically programmable optical random neural networks [0.0]
We demonstrate a genetically programmable yet simple optical neural network to achieve high performances with optical random projection.
By genetically programming the orientation of the scattering medium which acts as a random projection kernel, our novel technique finds an optimum kernel and improves its initial test accuracies 7-22%.
Our optical computing method presents a promising approach to achieve high performance in optical neural networks with a simple and scalable design.
arXiv Detail & Related papers (2024-03-19T06:55:59Z) - TeMPO: Efficient Time-Multiplexed Dynamic Photonic Tensor Core for Edge
AI with Compact Slow-Light Electro-Optic Modulator [44.74560543672329]
We present a time-multiplexed dynamic photonic tensor accelerator, dubbed TeMPO, with cross-layer device/circuit/architecture customization.
We achieve a 368.6 TOPS peak performance, 22.3 TOPS/W energy efficiency, and 1.2 TOPS/mm$2$ compute density.
This work signifies the power of cross-layer co-design and domain-specific customization, paving the way for future electronic-photonic accelerators.
arXiv Detail & Related papers (2024-02-12T03:40:32Z) - Spatially Varying Nanophotonic Neural Networks [39.1303097259564]
Photonic processors that execute operations using photons instead of electrons promise to enable optical neural networks with ultra-low latency and power consumption.
Existing optical neural networks, limited by the underlying network designs, have achieved image recognition accuracy far below that of state-of-the-art electronic neural networks.
arXiv Detail & Related papers (2023-08-07T08:48:46Z) - Quantum-noise-limited optical neural networks operating at a few quanta
per activation [5.494796517705931]
We show that it is possible to train optical neural networks to perform deterministic image-classification tasks with high accuracy.
We experimentally demonstrated MNIST classification with a test accuracy of 98% using an optical neural network with a hidden layer operating in the single-photon regime.
arXiv Detail & Related papers (2023-07-28T17:59:46Z) - Single-Shot Optical Neural Network [55.41644538483948]
'Weight-stationary' analog optical and electronic hardware has been proposed to reduce the compute resources required by deep neural networks.
We present a scalable, single-shot-per-layer weight-stationary optical processor.
arXiv Detail & Related papers (2022-05-18T17:49:49Z) - 11 TeraFLOPs per second photonic convolutional accelerator for deep
learning optical neural networks [0.0]
We demonstrate a universal optical vector convolutional accelerator operating beyond 10 TeraFLOPS (floating point operations per second)
We then use the same hardware to sequentially form a deep optical CNN with ten output neurons, achieving successful recognition of full 10 digits with 900 pixel handwritten digit images with 88% accuracy.
This approach is scalable and trainable to much more complex networks for demanding applications such as unmanned vehicle and real-time video recognition.
arXiv Detail & Related papers (2020-11-14T21:24:01Z) - Rapid characterisation of linear-optical networks via PhaseLift [51.03305009278831]
Integrated photonics offers great phase-stability and can rely on the large scale manufacturability provided by the semiconductor industry.
New devices, based on such optical circuits, hold the promise of faster and energy-efficient computations in machine learning applications.
We present a novel technique to reconstruct the transfer matrix of linear optical networks.
arXiv Detail & Related papers (2020-10-01T16:04:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.