Super-resolution image display using diffractive decoders
- URL: http://arxiv.org/abs/2206.07281v1
- Date: Wed, 15 Jun 2022 03:42:36 GMT
- Title: Super-resolution image display using diffractive decoders
- Authors: Cagatay Isil, Deniz Mengu, Yifan Zhao, Anika Tabassum, Jingxi Li, Yi
Luo, Mona Jarrahi, and Aydogan Ozcan
- Abstract summary: High-resolution synthesis/projection of images over a large field-of-view (FOV) is hindered by the restricted space-bandwidth-product (SBP) of wavefront modulators.
We report a deep learning-enabled diffractive display design that is based on a jointly-trained pair of an electronic encoder and a diffractive optical decoder.
Our results indicate that this diffractive image display can achieve a super-resolution factor of 4, demonstrating a 16-fold increase in SBP.
- Score: 21.24387597787123
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: High-resolution synthesis/projection of images over a large field-of-view
(FOV) is hindered by the restricted space-bandwidth-product (SBP) of wavefront
modulators. We report a deep learning-enabled diffractive display design that
is based on a jointly-trained pair of an electronic encoder and a diffractive
optical decoder to synthesize/project super-resolved images using
low-resolution wavefront modulators. The digital encoder, composed of a trained
convolutional neural network (CNN), rapidly pre-processes the high-resolution
images of interest so that their spatial information is encoded into
low-resolution (LR) modulation patterns, projected via a low SBP wavefront
modulator. The diffractive decoder processes this LR encoded information using
thin transmissive layers that are structured using deep learning to
all-optically synthesize and project super-resolved images at its output FOV.
Our results indicate that this diffractive image display can achieve a
super-resolution factor of ~4, demonstrating a ~16-fold increase in SBP. We
also experimentally validate the success of this diffractive super-resolution
display using 3D-printed diffractive decoders that operate at the THz spectrum.
This diffractive image decoder can be scaled to operate at visible wavelengths
and inspire the design of large FOV and high-resolution displays that are
compact, low-power, and computationally efficient.
Related papers
- Pixel-Aligned Multi-View Generation with Depth Guided Decoder [86.1813201212539]
We propose a novel method for pixel-level image-to-multi-view generation.
Unlike prior work, we incorporate attention layers across multi-view images in the VAE decoder of a latent video diffusion model.
Our model enables better pixel alignment across multi-view images.
arXiv Detail & Related papers (2024-08-26T04:56:41Z) - Learning Spatial Adaptation and Temporal Coherence in Diffusion Models for Video Super-Resolution [151.1255837803585]
We propose a novel approach, pursuing Spatial Adaptation and Temporal Coherence (SATeCo) for video super-resolution.
SATeCo pivots on learning spatial-temporal guidance from low-resolution videos to calibrate both latent-space high-resolution video denoising and pixel-space video reconstruction.
Experiments conducted on the REDS4 and Vid4 datasets demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2024-03-25T17:59:26Z) - Subwavelength Imaging using a Solid-Immersion Diffractive Optical
Processor [9.47970290529295]
We develop a compact, all-optical diffractive imager for subwavelength imaging of phase objects.
The imager can find wide-ranging applications in bioimaging, endoscopy, sensing and materials characterization.
arXiv Detail & Related papers (2024-01-17T02:12:57Z) - More complex encoder is not all you need [0.882348769487259]
We introduce neU-Net (i.e., not complex encoder U-Net), which incorporates a novel Sub-pixel Convolution for upsampling to construct a powerful decoder.
Our model design achieves excellent results, surpassing other state-of-the-art methods on both the Synapse and ACDC datasets.
arXiv Detail & Related papers (2023-09-20T08:34:38Z) - Pyramid diffractive optical networks for unidirectional image magnification and demagnification [0.0]
We present a pyramid-structured diffractive optical network design (which we term P-D2NN) for unidirectional image magnification and demagnification.
The P-D2NN design creates high-fidelity magnified or demagnified images in only one direction, while inhibiting the image formation in the opposite direction.
arXiv Detail & Related papers (2023-08-29T04:46:52Z) - Learning Kernel-Modulated Neural Representation for Efficient Light
Field Compression [41.24757573290883]
We design a compact neural network representation for the light field compression task.
It is composed of two types of complementary kernels: descriptive kernels (descriptors) that store scene description information learned during training, and modulatory kernels (modulators) that control the rendering of different SAIs from the queried perspectives.
arXiv Detail & Related papers (2023-07-12T12:58:03Z) - Neural Data-Dependent Transform for Learned Image Compression [72.86505042102155]
We build a neural data-dependent transform and introduce a continuous online mode decision mechanism to jointly optimize the coding efficiency for each individual image.
The experimental results show the effectiveness of the proposed neural-syntax design and the continuous online mode decision mechanism.
arXiv Detail & Related papers (2022-03-09T14:56:48Z) - Reducing Redundancy in the Bottleneck Representation of the Autoencoders [98.78384185493624]
Autoencoders are a type of unsupervised neural networks, which can be used to solve various tasks.
We propose a scheme to explicitly penalize feature redundancies in the bottleneck representation.
We tested our approach across different tasks: dimensionality reduction using three different dataset, image compression using the MNIST dataset, and image denoising using fashion MNIST.
arXiv Detail & Related papers (2022-02-09T18:48:02Z) - Neural Étendue Expander for Ultra-Wide-Angle High-Fidelity Holographic Display [51.399291206537384]
Modern holographic displays possess low 'etendue, which is the product of the display area and the maximum solid angle of diffracted light.
We present neural 'etendue expanders, which are learned from a natural image dataset.
With neural 'etendue expanders, we experimentally achieve 64$times$ 'etendue expansion of natural images in full color, expanding the FOV by an order of magnitude horizontally and vertically.
arXiv Detail & Related papers (2021-09-16T17:21:52Z) - A Novel Unified Model for Multi-exposure Stereo Coding Based on Low Rank
Tucker-ALS and 3D-HEVC [0.6091702876917279]
We propose an efficient scheme for coding multi-exposure stereo images based on a tensor low-rank approximation scheme.
The multi-exposure fusion can be realized to generate HDR stereo output at the decoder for increased realism and binocular 3D depth cues.
The encoding with 3D-HEVC enhance the proposed scheme efficiency by exploiting intra-frame, inter-view and the inter-component redundancies in lowrank approximated representation.
arXiv Detail & Related papers (2021-04-10T10:10:14Z) - Learned Multi-View Texture Super-Resolution [76.82725815863711]
We present a super-resolution method capable of creating a high-resolution texture map for a virtual 3D object from a set of lower-resolution images of that object.
Our architecture unifies the concepts of (i) multi-view super-resolution based on the redundancy of overlapping views and (ii) single-view super-resolution based on a learned prior of high-resolution (HR) image structure.
arXiv Detail & Related papers (2020-01-14T13:49:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.