On the interplay between physical and content priors in deep learning
for computational imaging
- URL: http://arxiv.org/abs/2004.06355v1
- Date: Tue, 14 Apr 2020 08:36:46 GMT
- Title: On the interplay between physical and content priors in deep learning
for computational imaging
- Authors: Mo Deng, Shuai Li, Iksung Kang, Nicholas X. Fang and George
Barbastathis
- Abstract summary: We use the Phase Extraction Neural Network (PhENN) for quantitative phase retrieval in a lensless phase imaging system.
We show that the two questions are related and share a common crux: the choice of the training examples.
We also discover that weaker regularization effect leads to better learning of the underlying propagation model.
- Score: 5.486833154281385
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning (DL) has been applied extensively in many computational imaging
problems, often leading to superior performance over traditional iterative
approaches. However, two important questions remain largely unanswered: first,
how well can the trained neural network generalize to objects very different
from the ones in training? This is particularly important in practice, since
large-scale annotated examples similar to those of interest are often not
available during training. Second, has the trained neural network learnt the
underlying (inverse) physics model, or has it merely done something trivial,
such as memorizing the examples or point-wise pattern matching? This pertains
to the interpretability of machine-learning based algorithms. In this work, we
use the Phase Extraction Neural Network (PhENN), a deep neural network (DNN)
for quantitative phase retrieval in a lensless phase imaging system as the
standard platform and show that the two questions are related and share a
common crux: the choice of the training examples. Moreover, we connect the
strength of the regularization effect imposed by a training set to the training
process with the Shannon entropy of images in the dataset. That is, the higher
the entropy of the training images, the weaker the regularization effect can be
imposed. We also discover that weaker regularization effect leads to better
learning of the underlying propagation model, i.e. the weak object transfer
function, applicable for weakly scattering objects under the weak object
approximation. Finally, simulation and experimental results show that better
cross-domain generalization performance can be achieved if DNN is trained on a
higher-entropy database, e.g. the ImageNet, than if the same DNN is trained on
a lower-entropy database, e.g. MNIST, as the former allows the underlying
physics model be learned better than the latter.
Related papers
- Neural networks trained with SGD learn distributions of increasing
complexity [78.30235086565388]
We show that neural networks trained using gradient descent initially classify their inputs using lower-order input statistics.
We then exploit higher-order statistics only later during training.
We discuss the relation of DSB to other simplicity biases and consider its implications for the principle of universality in learning.
arXiv Detail & Related papers (2022-11-21T15:27:22Z) - EfficientTrain: Exploring Generalized Curriculum Learning for Training
Visual Backbones [80.662250618795]
This paper presents a new curriculum learning approach for the efficient training of visual backbones (e.g., vision Transformers)
As an off-the-shelf method, it reduces the wall-time training cost of a wide variety of popular models by >1.5x on ImageNet-1K/22K without sacrificing accuracy.
arXiv Detail & Related papers (2022-11-17T17:38:55Z) - Learning sparse features can lead to overfitting in neural networks [9.2104922520782]
We show that feature learning can perform worse than lazy training.
Although sparsity is known to be essential for learning anisotropic data, it is detrimental when the target function is constant or smooth.
arXiv Detail & Related papers (2022-06-24T14:26:33Z) - Neural Maximum A Posteriori Estimation on Unpaired Data for Motion
Deblurring [87.97330195531029]
We propose a Neural Maximum A Posteriori (NeurMAP) estimation framework for training neural networks to recover blind motion information and sharp content from unpaired data.
The proposed NeurMAP is an approach to existing deblurring neural networks, and is the first framework that enables training image deblurring networks on unpaired datasets.
arXiv Detail & Related papers (2022-04-26T08:09:47Z) - Inverse-Dirichlet Weighting Enables Reliable Training of Physics
Informed Neural Networks [2.580765958706854]
We describe and remedy a failure mode that may arise from multi-scale dynamics with scale imbalances during training of deep neural networks.
PINNs are popular machine-learning templates that allow for seamless integration of physical equation models with data.
For inverse modeling using sequential training, we find that inverse-Dirichlet weighting protects a PINN against catastrophic forgetting.
arXiv Detail & Related papers (2021-07-02T10:01:37Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z) - Retrieval Augmentation to Improve Robustness and Interpretability of
Deep Neural Networks [3.0410237490041805]
In this work, we actively exploit the training data to improve the robustness and interpretability of deep neural networks.
Specifically, the proposed approach uses the target of the nearest input example to initialize the memory state of an LSTM model or to guide attention mechanisms.
Results show the effectiveness of the proposed models for the two tasks, on the widely used Flickr8 and IMDB datasets.
arXiv Detail & Related papers (2021-02-25T17:38:31Z) - The Surprising Simplicity of the Early-Time Learning Dynamics of Neural
Networks [43.860358308049044]
In work, we show that these common perceptions can be completely false in the early phase of learning.
We argue that this surprising simplicity can persist in networks with more layers with convolutional architecture.
arXiv Detail & Related papers (2020-06-25T17:42:49Z) - BP-DIP: A Backprojection based Deep Image Prior [49.375539602228415]
We propose two image restoration approaches: (i) Deep Image Prior (DIP), which trains a convolutional neural network (CNN) from scratch in test time using the degraded image; and (ii) a backprojection (BP) fidelity term, which is an alternative to the standard least squares loss that is usually used in previous DIP works.
We demonstrate the performance of the proposed method, termed BP-DIP, on the deblurring task and show its advantages over the plain DIP, with both higher PSNR values and better inference run-time.
arXiv Detail & Related papers (2020-03-11T17:09:12Z) - The large learning rate phase of deep learning: the catapult mechanism [50.23041928811575]
We present a class of neural networks with solvable training dynamics.
We find good agreement between our model's predictions and training dynamics in realistic deep learning settings.
We believe our results shed light on characteristics of models trained at different learning rates.
arXiv Detail & Related papers (2020-03-04T17:52:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.