Optical multi-task learning using multi-wavelength diffractive deep
neural networks
- URL: http://arxiv.org/abs/2212.00022v1
- Date: Wed, 30 Nov 2022 14:27:14 GMT
- Title: Optical multi-task learning using multi-wavelength diffractive deep
neural networks
- Authors: Zhengyang Duan, Hang Chen, Xing Lin
- Abstract summary: Photonic neural networks are brain-inspired information processing technology using photons instead of electrons to perform AI tasks.
Existing architectures are designed for a single task but fail to multiplex different tasks in parallel within a single monolithic system.
This paper proposes a novel optical multi-task learning system by designing multi-wavelength diffractive deep neural networks (D2NNs) with the joint optimization method.
- Score: 8.543496127018567
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Photonic neural networks are brain-inspired information processing technology
using photons instead of electrons to perform artificial intelligence (AI)
tasks. However, existing architectures are designed for a single task but fail
to multiplex different tasks in parallel within a single monolithic system due
to the task competition that deteriorates the model performance. This paper
proposes a novel optical multi-task learning system by designing
multi-wavelength diffractive deep neural networks (D2NNs) with the joint
optimization method. By encoding multi-task inputs into multi-wavelength
channels, the system can increase the computing throughput and significantly
alle-viate the competition to perform multiple tasks in parallel with high
accuracy. We design the two-task and four-task D2NNs with two and four spectral
channels, respectively, for classifying different inputs from MNIST, FMNIST,
KMNIST, and EMNIST databases. The numerical evaluations demonstrate that, under
the same network size, mul-ti-wavelength D2NNs achieve significantly higher
classification accuracies for multi-task learning than single-wavelength D2NNs.
Furthermore, by increasing the network size, the multi-wavelength D2NNs for
simultaneously performing multiple tasks achieve comparable classification
accuracies with respect to the individual training of multiple
single-wavelength D2NNs to perform tasks separately. Our work paves the way for
developing the wave-length-division multiplexing technology to achieve
high-throughput neuromorphic photonic computing and more general AI systems to
perform multiple tasks in parallel.
Related papers
- Neuromorphic Wireless Split Computing with Multi-Level Spikes [69.73249913506042]
In neuromorphic computing, spiking neural networks (SNNs) perform inference tasks, offering significant efficiency gains for workloads involving sequential data.
Recent advances in hardware and software have demonstrated that embedding a few bits of payload in each spike exchanged between the spiking neurons can further enhance inference accuracy.
This paper investigates a wireless neuromorphic split computing architecture employing multi-level SNNs.
arXiv Detail & Related papers (2024-11-07T14:08:35Z) - Provable Multi-Task Representation Learning by Two-Layer ReLU Neural Networks [69.38572074372392]
We present the first results proving that feature learning occurs during training with a nonlinear model on multiple tasks.
Our key insight is that multi-task pretraining induces a pseudo-contrastive loss that favors representations that align points that typically have the same label across tasks.
arXiv Detail & Related papers (2023-07-13T16:39:08Z) - Fast GraspNeXt: A Fast Self-Attention Neural Network Architecture for
Multi-task Learning in Computer Vision Tasks for Robotic Grasping on the Edge [80.88063189896718]
High architectural and computational complexity can result in poor suitability for deployment on embedded devices.
Fast GraspNeXt is a fast self-attention neural network architecture tailored for embedded multi-task learning in computer vision tasks for robotic grasping.
arXiv Detail & Related papers (2023-04-21T18:07:14Z) - OFA$^2$: A Multi-Objective Perspective for the Once-for-All Neural
Architecture Search [79.36688444492405]
Once-for-All (OFA) is a Neural Architecture Search (NAS) framework designed to address the problem of searching efficient architectures for devices with different resources constraints.
We aim to give one step further in the search for efficiency by explicitly conceiving the search stage as a multi-objective optimization problem.
arXiv Detail & Related papers (2023-03-23T21:30:29Z) - DeepSeq: Deep Sequential Circuit Learning [10.402436619244911]
Circuit representation learning is a promising research direction in the electronic design automation (EDA) field.
Existing solutions only target combinational circuits, significantly limiting their applications.
We propose DeepSeq, a novel representation learning framework for sequential netlists.
arXiv Detail & Related papers (2023-02-27T09:17:35Z) - L-HYDRA: Multi-Head Physics-Informed Neural Networks [0.0]
We construct multi-head physics-informed neural networks (MH-PINNs) as potent tool for multi-task learning (MTL), generative modeling, and few-shot learning.
MH-PINNs connect multiple functions/tasks via a shared body as the basis functions as well as a shared distribution for the head.
We demonstrate the effectiveness of MH-PINNs in five benchmarks, investigating also the possibility of synergistic learning in regression analysis.
arXiv Detail & Related papers (2023-01-05T16:54:01Z) - M$^3$ViT: Mixture-of-Experts Vision Transformer for Efficient Multi-task
Learning with Model-Accelerator Co-design [95.41238363769892]
Multi-task learning (MTL) encapsulates multiple learned tasks in a single model and often lets those tasks learn better jointly.
Current MTL regimes have to activate nearly the entire model even to just execute a single task.
We present a model-accelerator co-design framework to enable efficient on-device MTL.
arXiv Detail & Related papers (2022-10-26T15:40:24Z) - MT-SNN: Spiking Neural Network that Enables Single-Tasking of Multiple
Tasks [0.0]
We implement a multi-task spiking neural network (MT-SNN) that can learn two or more classification tasks while performing one task at a time.
The network is implemented using Intel's Lava platform for the Loihi2 neuromorphic chip.
arXiv Detail & Related papers (2022-08-02T15:17:07Z) - Real-time Multi-Task Diffractive Deep Neural Networks via
Hardware-Software Co-design [1.6066483376871004]
This work proposes a novel hardware-software co-design method that enables robust and noise-resilient Multi-task Learning in D$2$NNs.
Our experimental results demonstrate significant improvements in versatility and hardware efficiency, and also demonstrate the robustness of proposed multi-task D$2$NN architecture.
arXiv Detail & Related papers (2020-12-16T12:29:54Z) - Deep Multimodal Neural Architecture Search [178.35131768344246]
We devise a generalized deep multimodal neural architecture search (MMnas) framework for various multimodal learning tasks.
Given multimodal input, we first define a set of primitive operations, and then construct a deep encoder-decoder based unified backbone.
On top of the unified backbone, we attach task-specific heads to tackle different multimodal learning tasks.
arXiv Detail & Related papers (2020-04-25T07:00:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.