Pyramid diffractive optical networks for unidirectional magnification
and demagnification
- URL: http://arxiv.org/abs/2308.15019v1
- Date: Tue, 29 Aug 2023 04:46:52 GMT
- Title: Pyramid diffractive optical networks for unidirectional magnification
and demagnification
- Authors: Bijie Bai, Xilin Yang, Tianyi Gan, Jingxi Li, Deniz Mengu, Mona
Jarrahi, Aydogan Ozcan
- Abstract summary: We present a pyramid-structured diffractive optical network design (which we term P-D2NN) for unidirectional image magnification and demagnification.
Our analyses revealed the efficacy of this P-D2NN design in unidirectional image magnification and demagnification tasks.
It produces high-fidelity magnified or demagnified images in only one direction, while inhibiting the image formation in the opposite direction.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Diffractive deep neural networks (D2NNs) are composed of successive
transmissive layers optimized using supervised deep learning to all-optically
implement various computational tasks between an input and output field-of-view
(FOV). Here, we present a pyramid-structured diffractive optical network design
(which we term P-D2NN), optimized specifically for unidirectional image
magnification and demagnification. In this P-D2NN design, the diffractive
layers are pyramidally scaled in alignment with the direction of the image
magnification or demagnification. Our analyses revealed the efficacy of this
P-D2NN design in unidirectional image magnification and demagnification tasks,
producing high-fidelity magnified or demagnified images in only one direction,
while inhibiting the image formation in the opposite direction - confirming the
desired unidirectional imaging operation. Compared to the conventional D2NN
designs with uniform-sized successive diffractive layers, P-D2NN design
achieves similar performance in unidirectional magnification tasks using only
half of the diffractive degrees of freedom within the optical processor volume.
Furthermore, it maintains its unidirectional image
magnification/demagnification functionality across a large band of illumination
wavelengths despite being trained with a single illumination wavelength. With
this pyramidal architecture, we also designed a wavelength-multiplexed
diffractive network, where a unidirectional magnifier and a unidirectional
demagnifier operate simultaneously in opposite directions, at two distinct
illumination wavelengths. The efficacy of the P-D2NN architecture was also
validated experimentally using monochromatic terahertz illumination,
successfully matching our numerical simulations. P-D2NN offers a
physics-inspired strategy for designing task-specific visual processors.
Related papers
- 2-D SSM: A General Spatial Layer for Visual Transformers [79.4957965474334]
A central objective in computer vision is to design models with appropriate 2-D inductive bias.
We leverage an expressive variation of the multidimensional State Space Model.
Our approach introduces efficient parameterization, accelerated computation, and a suitable normalization scheme.
arXiv Detail & Related papers (2023-06-11T09:41:37Z) - SAR-to-Optical Image Translation via Thermodynamics-inspired Network [68.71771171637677]
A Thermodynamics-inspired Network for SAR-to-Optical Image Translation (S2O-TDN) is proposed in this paper.
S2O-TDN follows an explicit design principle derived from thermodynamic theory and enjoys the advantage of explainability.
Experiments on the public SEN1-2 dataset show the advantages of the proposed S2O-TDN over the current methods with more delicate textures and higher quantitative results.
arXiv Detail & Related papers (2023-05-23T09:02:33Z) - Unidirectional Imaging using Deep Learning-Designed Materials [13.048762595058058]
A unidirectional imager would only permit image formation along one direction, from an input field-of-view (FOV) A to an output FOV B, and in the reverse path.
Here, we report the first demonstration of unidirectional imagers, presenting polarization-insensitive and broadband unidirectional imaging based on successive diffractive layers that are linear and isotropic.
These diffractive layers are optimized using deep learning and consist of hundreds of thousands of diffractive phase features, which collectively modulate the incoming fields and project an intensity image of the input onto an output FOV, while blocking the image formation in the
arXiv Detail & Related papers (2022-12-05T04:43:03Z) - Time-lapse image classification using a diffractive neural network [0.0]
We show for the first time a time-lapse image classification scheme using a diffractive network.
We show a blind testing accuracy of 62.03% on the optical classification of objects from the CIFAR-10 dataset.
This constitutes the highest inference accuracy achieved so far using a single diffractive network.
arXiv Detail & Related papers (2022-08-23T08:16:30Z) - D$^\text{2}$UF: Deep Coded Aperture Design and Unrolling Algorithm for
Compressive Spectral Image Fusion [22.0246327137227]
This paper presents the fusion of the compressive measurements of a low-spatial high-spectral resolution coded aperture snapshot spectral imager (CASSI) architecture and a high-spatial low-spectral resolution multispectral color filter array (MCFA) system.
Unlike previous CSIF works, this paper proposes joint optimization of the sensing architectures and a reconstruction network in an end-to-end (E2E) manner.
arXiv Detail & Related papers (2022-05-24T15:39:34Z) - All-optical graph representation learning using integrated diffractive
photonic computing units [51.15389025760809]
Photonic neural networks perform brain-inspired computations using photons instead of electrons.
We propose an all-optical graph representation learning architecture, termed diffractive graph neural network (DGNN)
We demonstrate the use of DGNN extracted features for node and graph-level classification tasks with benchmark databases and achieve superior performance.
arXiv Detail & Related papers (2022-04-23T02:29:48Z) - Disentangling Light Fields for Super-Resolution and Disparity Estimation [67.50796924758221]
Light field (LF) cameras record both intensity and directions of light rays, and encode 3D scenes into 4D LF images.
It is challenging for convolutional neural networks (CNNs) to process LF images since the spatial and angular information are highly inter-twined with varying disparities.
We propose a generic mechanism to disentangle these coupled information for LF image processing.
arXiv Detail & Related papers (2022-02-22T01:04:41Z) - High-Resolution Optical Flow from 1D Attention and Correlation [89.61824964952949]
We propose a new method for high-resolution optical flow estimation with significantly less computation.
We first perform a 1D attention operation in the vertical direction of the target image, and then a simple 1D correlation in the horizontal direction of the attended image.
Experiments on Sintel, KITTI and real-world 4K resolution images demonstrated the effectiveness and superiority of our proposed method.
arXiv Detail & Related papers (2021-04-28T17:56:34Z) - Leveraging Spatial and Photometric Context for Calibrated Non-Lambertian
Photometric Stereo [61.6260594326246]
We introduce an efficient fully-convolutional architecture that can leverage both spatial and photometric context simultaneously.
Using separable 4D convolutions and 2D heat-maps reduces the size and makes more efficient.
arXiv Detail & Related papers (2021-03-22T18:06:58Z) - Learning Inter- and Intra-frame Representations for Non-Lambertian
Photometric Stereo [14.5172791293107]
We build a two-stage Convolutional Neural Network (CNN) architecture to construct inter- and intra-frame representations.
We experimentally investigate numerous network design alternatives for identifying the optimal scheme to deploy inter-frame and intra-frame feature extraction modules.
arXiv Detail & Related papers (2020-12-26T11:22:56Z) - Deep learning-based virtual refocusing of images using an engineered
point-spread function [1.2977570993112095]
We present a virtual image refocusing method over an extended depth of field (DOF) enabled by cascaded neural networks and a double-helix point-spread function (DH-PSF)
We extend the DOF of a fluorescence microscope by 20-fold.
arXiv Detail & Related papers (2020-12-22T09:15:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.