Data-efficient Modeling of Optical Matrix Multipliers Using Transfer
Learning
- URL: http://arxiv.org/abs/2211.16038v1
- Date: Tue, 29 Nov 2022 09:22:42 GMT
- Title: Data-efficient Modeling of Optical Matrix Multipliers Using Transfer
Learning
- Authors: Ali Cem, Ognjen Jovanovic, Siqi Yan, Yunhong Ding, Darko Zibar,
Francesco Da Ros
- Abstract summary: We demonstrate transfer learning-assisted neural network models for optical matrix multipliers with scarce measurement data.
Our approach uses 10% of experimental data needed for best performance and outperforms analytical models for a Mach-Zehnder interferometer mesh.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We demonstrate transfer learning-assisted neural network models for optical
matrix multipliers with scarce measurement data. Our approach uses <10\% of
experimental data needed for best performance and outperforms analytical models
for a Mach-Zehnder interferometer mesh.
Related papers
- Addressing Data Scarcity in Optical Matrix Multiplier Modeling Using
Transfer Learning [0.0]
We present and experimentally evaluate using transfer learning to address experimental data scarcity.
Our approach involves pre-training the model using synthetic data generated from a less accurate analytical model.
We achieve 1 dB root-mean-square error on the matrix weights implemented by a 3x3 photonic chip while using only 25% of the available data.
arXiv Detail & Related papers (2023-08-10T07:33:00Z) - Machine learning enabled experimental design and parameter estimation
for ultrafast spin dynamics [54.172707311728885]
We introduce a methodology that combines machine learning with Bayesian optimal experimental design (BOED)
Our method employs a neural network model for large-scale spin dynamics simulations for precise distribution and utility calculations in BOED.
Our numerical benchmarks demonstrate the superior performance of our method in guiding XPFS experiments, predicting model parameters, and yielding more informative measurements within limited experimental time.
arXiv Detail & Related papers (2023-06-03T06:19:20Z) - Post-training Model Quantization Using GANs for Synthetic Data
Generation [57.40733249681334]
We investigate the use of synthetic data as a substitute for the calibration with real data for the quantization method.
We compare the performance of models quantized using data generated by StyleGAN2-ADA and our pre-trained DiStyleGAN, with quantization using real data and an alternative data generation method based on fractal images.
arXiv Detail & Related papers (2023-05-10T11:10:09Z) - An Adversarial Active Sampling-based Data Augmentation Framework for
Manufacturable Chip Design [55.62660894625669]
Lithography modeling is a crucial problem in chip design to ensure a chip design mask is manufacturable.
Recent developments in machine learning have provided alternative solutions in replacing the time-consuming lithography simulations with deep neural networks.
We propose a litho-aware data augmentation framework to resolve the dilemma of limited data and improve the machine learning model performance.
arXiv Detail & Related papers (2022-10-27T20:53:39Z) - Data-driven Modeling of Mach-Zehnder Interferometer-based Optical Matrix
Multipliers [0.0]
Photonic integrated circuits are facilitating the development of optical neural networks.
We describe both simple analytical models and data-driven models for offline training of optical matrix multipliers.
The neural network-based models outperform the simple physics-based models in terms of prediction error.
arXiv Detail & Related papers (2022-10-17T15:19:26Z) - Mixed Effects Neural ODE: A Variational Approximation for Analyzing the
Dynamics of Panel Data [50.23363975709122]
We propose a probabilistic model called ME-NODE to incorporate (fixed + random) mixed effects for analyzing panel data.
We show that our model can be derived using smooth approximations of SDEs provided by the Wong-Zakai theorem.
We then derive Evidence Based Lower Bounds for ME-NODE, and develop (efficient) training algorithms.
arXiv Detail & Related papers (2022-02-18T22:41:51Z) - Comparison of Models for Training Optical Matrix Multipliers in
Neuromorphic PICs [0.0]
We compare simple physics-based vs. data-driven neural-network-based models for offline training of programmable photonic chips.
The neural-network model outperforms physics-based models for a chip with thermal crosstalk, yielding increased testing accuracy.
arXiv Detail & Related papers (2021-11-23T12:15:21Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z) - A Taylor Based Sampling Scheme for Machine Learning in Computational
Physics [0.0]
We take advantage of the ability to generate data using numerical simulations programs to train Machine Learning models better.
We elaborate a new data sampling scheme based on Taylor approximation to reduce the error of a Deep Neural Network (DNN) when learning the solution of an ordinary differential equations (ODE) system.
arXiv Detail & Related papers (2021-01-20T12:56:09Z) - Transfer Learning without Knowing: Reprogramming Black-box Machine
Learning Models with Scarce Data and Limited Resources [78.72922528736011]
We propose a novel approach, black-box adversarial reprogramming (BAR), that repurposes a well-trained black-box machine learning model.
Using zeroth order optimization and multi-label mapping techniques, BAR can reprogram a black-box ML model solely based on its input-output responses.
BAR outperforms state-of-the-art methods and yields comparable performance to the vanilla adversarial reprogramming method.
arXiv Detail & Related papers (2020-07-17T01:52:34Z) - Exponentially improved detection and correction of errors in
experimental systems using neural networks [0.0]
We introduce the use of two machine learning algorithms to create an empirical model of an experimental apparatus.
This is able to reduce the number of measurements necessary for generic optimisation tasks exponentially.
We demonstrate both algorithms at the example of detecting and compensating stray electric fields in an ion trap.
arXiv Detail & Related papers (2020-05-18T22:42:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.