Physics-aware Roughness Optimization for Diffractive Optical Neural
Networks
- URL: http://arxiv.org/abs/2304.01500v1
- Date: Tue, 4 Apr 2023 03:19:36 GMT
- Title: Physics-aware Roughness Optimization for Diffractive Optical Neural
Networks
- Authors: Shanglin Zhou, Yingjie Li, Minhan Lou, Weilu Gao, Zhijie Shi, Cunxi
Yu, Caiwen Ding
- Abstract summary: diffractive optical neural networks (DONNs) have shown promising advantages over conventional deep neural networks.
We propose a physics-aware diffractive optical neural network training framework to reduce the performance difference between numerical modeling and practical deployment.
- Score: 15.397285424104469
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: As a representative next-generation device/circuit technology beyond CMOS,
diffractive optical neural networks (DONNs) have shown promising advantages
over conventional deep neural networks due to extreme fast computation speed
(light speed) and low energy consumption. However, there is a mismatch, i.e.,
significant prediction accuracy loss, between the DONN numerical modelling and
physical optical device deployment, because of the interpixel interaction
within the diffractive layers. In this work, we propose a physics-aware
diffractive optical neural network training framework to reduce the performance
difference between numerical modeling and practical deployment. Specifically,
we propose the roughness modeling regularization in the training process and
integrate the physics-aware sparsification method to introduce sparsity to the
phase masks to reduce sharp phase changes between adjacent pixels in
diffractive layers. We further develop $2\pi$ periodic optimization to reduce
the roughness of the phase masks to preserve the performance of DONN.
Experiment results demonstrate that, compared to state-of-the-arts, our
physics-aware optimization can provide $35.7\%$, $34.2\%$, $28.1\%$, and
$27.3\%$ reduction in roughness with only accuracy loss on MNIST, FMNIST,
KMNIST, and EMNIST, respectively.
Related papers
- 1-bit Quantized On-chip Hybrid Diffraction Neural Network Enabled by Authentic All-optical Fully-connected Architecture [4.594367761345624]
This study introduces the Hybrid Diffraction Neural Network (HDNN), a novel architecture that incorporates matrix multiplication into DNNs.
utilizing a singular phase modulation layer and an amplitude modulation layer, the trained neural network demonstrated remarkable accuracies of 96.39% and 89% in digit recognition tasks.
arXiv Detail & Related papers (2024-04-11T02:54:17Z) - Gradual Optimization Learning for Conformational Energy Minimization [69.36925478047682]
Gradual Optimization Learning Framework (GOLF) for energy minimization with neural networks significantly reduces the required additional data.
Our results demonstrate that the neural network trained with GOLF performs on par with the oracle on a benchmark of diverse drug-like molecules.
arXiv Detail & Related papers (2023-11-05T11:48:08Z) - Accelerating Deep Neural Networks via Semi-Structured Activation
Sparsity [0.0]
Exploiting sparsity in the network's feature maps is one of the ways to reduce its inference latency.
We propose a solution to induce semi-structured activation sparsity exploitable through minor runtime modifications.
Our approach yields a speed improvement of $1.25 times$ with a minimal accuracy drop of $1.1%$ for the ResNet18 model on the ImageNet dataset.
arXiv Detail & Related papers (2023-09-12T22:28:53Z) - Classification robustness to common optical aberrations [64.08840063305313]
This paper proposes OpticsBench, a benchmark for investigating robustness to realistic, practically relevant optical blur effects.
Experiments on ImageNet show that for a variety of different pre-trained DNNs, the performance varies strongly compared to disk-shaped kernels.
We show on ImageNet-100 with OpticsAugment that can be increased by using optical kernels as data augmentation.
arXiv Detail & Related papers (2023-08-29T08:36:00Z) - Rubik's Optical Neural Networks: Multi-task Learning with Physics-aware
Rotation Architecture [14.55785957623462]
This work presents a novel ONNs architecture, namely, textitRubikONNs, which utilizes the physical properties of optical systems to encode multiple feed-forward functions.
Our experimental results demonstrate more than 4$times$ improvements in energy and cost efficiency with marginal accuracy compared to the state-of-the-art approaches.
arXiv Detail & Related papers (2023-04-25T16:51:12Z) - Globally Optimal Training of Neural Networks with Threshold Activation
Functions [63.03759813952481]
We study weight decay regularized training problems of deep neural networks with threshold activations.
We derive a simplified convex optimization formulation when the dataset can be shattered at a certain layer of the network.
arXiv Detail & Related papers (2023-03-06T18:59:13Z) - Fast Exploration of the Impact of Precision Reduction on Spiking Neural
Networks [63.614519238823206]
Spiking Neural Networks (SNNs) are a practical choice when the target hardware reaches the edge of computing.
We employ an Interval Arithmetic (IA) model to develop an exploration methodology that takes advantage of the capability of such a model to propagate the approximation error.
arXiv Detail & Related papers (2022-11-22T15:08:05Z) - Physics-aware Differentiable Discrete Codesign for Diffractive Optical
Neural Networks [12.952987240366781]
This work proposes a novel device-to-system hardware-software codesign framework, which enables efficient training of Diffractive optical neural networks (DONNs)
Gumbel-Softmax is employed to enable differentiable discrete mapping from real-world device parameters into the forward function of DONNs.
The results have demonstrated that our proposed framework offers significant advantages over conventional quantization-based methods.
arXiv Detail & Related papers (2022-09-28T17:13:28Z) - Optical Neural Ordinary Differential Equations [44.97261923694945]
We propose the optical neural ordinary differential equations (ON-ODE) architecture that parameterizes the continuous dynamics of hidden layers with optical ODE solvers.
The ON-ODE comprises the PNNs followed by the photonic integrator and optical feedback loop, which can be configured to represent residual neural networks (ResNet) and recurrent neural networks with effectively reduced chip area occupancy.
arXiv Detail & Related papers (2022-09-26T04:04:02Z) - Single-Shot Optical Neural Network [55.41644538483948]
'Weight-stationary' analog optical and electronic hardware has been proposed to reduce the compute resources required by deep neural networks.
We present a scalable, single-shot-per-layer weight-stationary optical processor.
arXiv Detail & Related papers (2022-05-18T17:49:49Z) - Learning Frequency-aware Dynamic Network for Efficient Super-Resolution [56.98668484450857]
This paper explores a novel frequency-aware dynamic network for dividing the input into multiple parts according to its coefficients in the discrete cosine transform (DCT) domain.
In practice, the high-frequency part will be processed using expensive operations and the lower-frequency part is assigned with cheap operations to relieve the computation burden.
Experiments conducted on benchmark SISR models and datasets show that the frequency-aware dynamic network can be employed for various SISR neural architectures.
arXiv Detail & Related papers (2021-03-15T12:54:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.