A Two-step-training Deep Learning Framework for Real-time Computational
Imaging without Physics Priors
- URL: http://arxiv.org/abs/2001.03493v3
- Date: Tue, 18 Aug 2020 13:57:55 GMT
- Title: A Two-step-training Deep Learning Framework for Real-time Computational
Imaging without Physics Priors
- Authors: Ruibo Shang, Kevin Hoffer-Hawlik, Geoffrey P. Luke
- Abstract summary: We propose a two-step-training DL (TST-DL) framework for real-time computational imaging without physics priors.
First, a single fully-connected layer (FCL) is trained to directly learn the model.
Then, this FCL is fixed and fixed with an un-trained U-Net architecture for a second-step training to improve the output image fidelity.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning (DL) is a powerful tool in computational imaging for many
applications. A common strategy is to reconstruct a preliminary image as the
input of a neural network to achieve an optimized image. Usually, the
preliminary image is acquired with the prior knowledge of the imaging model.
One outstanding challenge, however, is the degree to which the actual imaging
model deviates from the assumed model. Model mismatches degrade the quality of
the preliminary image and therefore affect the DL predictions. Another main
challenge is that since most imaging inverse problems are ill-posed and the
networks are over-parameterized, DL networks have flexibility to extract
features from the data that are not directly related to the imaging model. To
solve these challenges, a two-step-training DL (TST-DL) framework is proposed
for real-time computational imaging without physics priors. First, a single
fully-connected layer (FCL) is trained to directly learn the model. Then, this
FCL is fixed and concatenated with an un-trained U-Net architecture for a
second-step training to improve the output image fidelity, resulting in four
main advantages. First, it does not rely on an accurate representation of the
imaging model since the model is directly learned. Second, real-time imaging
can be achieved. Third, the TST-DL network is trained in the desired direction
and the predictions are improved since the first step is constrained to learn
the model and the second step improves the result by learning the optimal
regularizer. Fourth, the approach accommodates any size and dimensionality of
data. We demonstrate this framework using a linear single-pixel camera imaging
model. The results are quantitatively compared with those from other DL
frameworks and model-based iterative optimization approaches. We further extend
this concept to nonlinear models in the application of image
de-autocorrelation.
Related papers
- Fast constrained sampling in pre-trained diffusion models [77.21486516041391]
Diffusion models have dominated the field of large, generative image models.
We propose an algorithm for fast-constrained sampling in large pre-trained diffusion models.
arXiv Detail & Related papers (2024-10-24T14:52:38Z) - Calibrated Cache Model for Few-Shot Vision-Language Model Adaptation [36.45488536471859]
Similarity refines the image-image similarity by using unlabeled images.
Weight introduces a precision matrix into the weight function to adequately model the relation between training samples.
To reduce the high complexity of GPs, we propose a group-based learning strategy.
arXiv Detail & Related papers (2024-10-11T15:12:30Z) - A-SDM: Accelerating Stable Diffusion through Redundancy Removal and
Performance Optimization [54.113083217869516]
In this work, we first explore the computational redundancy part of the network.
We then prune the redundancy blocks of the model and maintain the network performance.
Thirdly, we propose a global-regional interactive (GRI) attention to speed up the computationally intensive attention part.
arXiv Detail & Related papers (2023-12-24T15:37:47Z) - Distance Weighted Trans Network for Image Completion [52.318730994423106]
We propose a new architecture that relies on Distance-based Weighted Transformer (DWT) to better understand the relationships between an image's components.
CNNs are used to augment the local texture information of coarse priors.
DWT blocks are used to recover certain coarse textures and coherent visual structures.
arXiv Detail & Related papers (2023-10-11T12:46:11Z) - P2P: Tuning Pre-trained Image Models for Point Cloud Analysis with
Point-to-Pixel Prompting [94.11915008006483]
We propose a novel Point-to-Pixel prompting for point cloud analysis.
Our method attains 89.3% accuracy on the hardest setting of ScanObjectNN.
Our framework also exhibits very competitive performance on ModelNet classification and ShapeNet Part Code.
arXiv Detail & Related papers (2022-08-04T17:59:03Z) - Stable Optimization for Large Vision Model Based Deep Image Prior in
Cone-Beam CT Reconstruction [6.558735319783205]
Large Vision Model (LVM) has recently demonstrated great potential for medical imaging tasks.
Deep Image Prior (DIP) effectively guides an untrained neural network to generate high-quality CBCT images without any training data.
We propose a stable optimization method for the forward-model-free DIP model for sparse-view CBCT.
arXiv Detail & Related papers (2022-03-23T15:16:29Z) - Meta Internal Learning [88.68276505511922]
Internal learning for single-image generation is a framework, where a generator is trained to produce novel images based on a single image.
We propose a meta-learning approach that enables training over a collection of images, in order to model the internal statistics of the sample image more effectively.
Our results show that the models obtained are as suitable as single-image GANs for many common image applications.
arXiv Detail & Related papers (2021-10-06T16:27:38Z) - Pre-Trained Image Processing Transformer [95.93031793337613]
We develop a new pre-trained model, namely, image processing transformer (IPT)
We present to utilize the well-known ImageNet benchmark for generating a large amount of corrupted image pairs.
IPT model is trained on these images with multi-heads and multi-tails.
arXiv Detail & Related papers (2020-12-01T09:42:46Z) - Efficient and Model-Based Infrared and Visible Image Fusion Via
Algorithm Unrolling [24.83209572888164]
Infrared and visible image fusion (IVIF) expects to obtain images that retain thermal radiation information from infrared images and texture details from visible images.
A model-based convolutional neural network (CNN) model is proposed to overcome the shortcomings of traditional CNN-based IVIF models.
arXiv Detail & Related papers (2020-05-12T16:15:56Z) - On the interplay between physical and content priors in deep learning
for computational imaging [5.486833154281385]
We use the Phase Extraction Neural Network (PhENN) for quantitative phase retrieval in a lensless phase imaging system.
We show that the two questions are related and share a common crux: the choice of the training examples.
We also discover that weaker regularization effect leads to better learning of the underlying propagation model.
arXiv Detail & Related papers (2020-04-14T08:36:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.