Transfer Learning in General Lensless Imaging through Scattering Media
- URL: http://arxiv.org/abs/1912.12419v1
- Date: Sat, 28 Dec 2019 07:37:25 GMT
- Title: Transfer Learning in General Lensless Imaging through Scattering Media
- Authors: Yukuan Yang, Lei Deng, Peng Jiao, Yansong Chua, Jing Pei, Cheng Ma,
Guoqi Li
- Abstract summary: Deep neural networks (DNNs) have been successfully introduced to the field of lensless imaging through scattering media.
In this work, transfer learning is proposed to address the problem of lensless imaging through scattering media.
- Score: 16.445963019768744
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently deep neural networks (DNNs) have been successfully introduced to the
field of lensless imaging through scattering media. By solving an inverse
problem in computational imaging, DNNs can overcome several shortcomings in the
conventional lensless imaging through scattering media methods, namely, high
cost, poor quality, complex control, and poor anti-interference. However, for
training, a large number of training samples on various datasets have to be
collected, with a DNN trained on one dataset generally performing poorly for
recovering images from another dataset. The underlying reason is that lensless
imaging through scattering media is a high dimensional regression problem and
it is difficult to obtain an analytical solution. In this work, transfer
learning is proposed to address this issue. Our main idea is to train a DNN on
a relatively complex dataset using a large number of training samples and
fine-tune the last few layers using very few samples from other datasets.
Instead of the thousands of samples required to train from scratch, transfer
learning alleviates the problem of costly data acquisition. Specifically,
considering the difference in sample sizes and similarity among datasets, we
propose two DNN architectures, namely LISMU-FCN and LISMU-OCN, and a balance
loss function designed for balancing smoothness and sharpness. LISMU-FCN, with
much fewer parameters, can achieve imaging across similar datasets while
LISMU-OCN can achieve imaging across significantly different datasets. What's
more, we establish a set of simulation algorithms which are close to the real
experiment, and it is of great significance and practical value in the research
on lensless scattering imaging. In summary, this work provides a new solution
for lensless imaging through scattering media using transfer learning in DNNs.
Related papers
- Analysis of Classifier Training on Synthetic Data for Cross-Domain Datasets [4.696575161583618]
This study focuses on camera-based traffic sign recognition applications for advanced driver assistance systems and autonomous driving.
The proposed augmentation pipeline of synthetic datasets includes novel augmentation processes such as structured shadows and gaussian specular highlights.
Experiments showed that a synthetic image-based approach outperforms in most cases real image-based training when applied to cross-domain test datasets.
arXiv Detail & Related papers (2024-10-30T07:11:41Z) - Deep Domain Adaptation: A Sim2Real Neural Approach for Improving Eye-Tracking Systems [80.62854148838359]
Eye image segmentation is a critical step in eye tracking that has great influence over the final gaze estimate.
We use dimensionality-reduction techniques to measure the overlap between the target eye images and synthetic training data.
Our methods result in robust, improved performance when tackling the discrepancy between simulation and real-world data samples.
arXiv Detail & Related papers (2024-03-23T22:32:06Z) - Impact of Light and Shadow on Robustness of Deep Neural Networks [5.015796849425367]
Deep neural networks (DNNs) have made remarkable strides in various computer vision tasks, including image classification, segmentation, and object detection.
Recent research has revealed a vulnerability in advanced DNNs when faced with deliberate manipulations of input data, known as adversarial attacks.
We propose a brightness-variation dataset, which incorporates 24 distinct brightness levels for each image within a subset of ImageNet.
arXiv Detail & Related papers (2023-05-23T15:30:56Z) - Robustifying the Multi-Scale Representation of Neural Radiance Fields [86.69338893753886]
We present a robust multi-scale neural radiance fields representation approach to overcome both real-world imaging issues.
Our method handles multi-scale imaging effects and camera-pose estimation problems with NeRF-inspired approaches.
We demonstrate, with examples, that for an accurate neural representation of an object from day-to-day acquired multi-view images, it is crucial to have precise camera-pose estimates.
arXiv Detail & Related papers (2022-10-09T11:46:45Z) - Agricultural Plantation Classification using Transfer Learning Approach
based on CNN [0.0]
The efficiency of hyper-spectral image recognition has increased significantly with deep learning.
CNN and Multi-Layer Perceptron(MLP) has demonstrated to be an excellent process of classifying images.
We propose using the method of transfer learning to decrease the training time and reduce the dependence on large labeled data-set.
arXiv Detail & Related papers (2022-06-19T14:43:31Z) - Terrain Classification using Transfer Learning on Hyperspectral Images:
A Comparative study [0.13999481573773068]
convolutional neural network (CNN) and the Multi-Layer Perceptron (MLP) have been proven to be an effective method of image classification.
However, they suffer from the issues of long training time and requirement of large amounts of the labeled data.
We propose using the method of transfer learning to decrease the training time and reduce the dependence on large labeled dataset.
arXiv Detail & Related papers (2022-06-19T14:36:33Z) - Feature transforms for image data augmentation [74.12025519234153]
In image classification, many augmentation approaches utilize simple image manipulation algorithms.
In this work, we build ensembles on the data level by adding images generated by combining fourteen augmentation approaches.
Pretrained ResNet50 networks are finetuned on training sets that include images derived from each augmentation method.
arXiv Detail & Related papers (2022-01-24T14:12:29Z) - Image Quality Assessment using Contrastive Learning [50.265638572116984]
We train a deep Convolutional Neural Network (CNN) using a contrastive pairwise objective to solve the auxiliary problem.
We show through extensive experiments that CONTRIQUE achieves competitive performance when compared to state-of-the-art NR image quality models.
Our results suggest that powerful quality representations with perceptual relevance can be obtained without requiring large labeled subjective image quality datasets.
arXiv Detail & Related papers (2021-10-25T21:01:00Z) - Deep Multimodal Transfer-Learned Regression in Data-Poor Domains [0.0]
We propose a Deep Multimodal Transfer-Learned Regressor (DMTL-R) for multimodal learning of image and feature data.
Our model is capable of fine-tuning a given set of pre-trained CNN weights on a small amount of training image data.
We present results using phase-field simulation microstructure images with an accompanying set of physical features, using pre-trained weights from various well-known CNN architectures.
arXiv Detail & Related papers (2020-06-16T16:52:44Z) - Data Consistent CT Reconstruction from Insufficient Data with Learned
Prior Images [70.13735569016752]
We investigate the robustness of deep learning in CT image reconstruction by showing false negative and false positive lesion cases.
We propose a data consistent reconstruction (DCR) method to improve their image quality, which combines the advantages of compressed sensing and deep learning.
The efficacy of the proposed method is demonstrated in cone-beam CT with truncated data, limited-angle data and sparse-view data, respectively.
arXiv Detail & Related papers (2020-05-20T13:30:49Z) - Single-Image HDR Reconstruction by Learning to Reverse the Camera
Pipeline [100.5353614588565]
We propose to incorporate the domain knowledge of the LDR image formation pipeline into our model.
We model the HDRto-LDR image formation pipeline as the (1) dynamic range clipping, (2) non-linear mapping from a camera response function, and (3) quantization.
We demonstrate that the proposed method performs favorably against state-of-the-art single-image HDR reconstruction algorithms.
arXiv Detail & Related papers (2020-04-02T17:59:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.