Learning Regularization Parameters of Inverse Problems via Deep Neural
Networks
- URL: http://arxiv.org/abs/2104.06594v1
- Date: Wed, 14 Apr 2021 02:38:38 GMT
- Title: Learning Regularization Parameters of Inverse Problems via Deep Neural
Networks
- Authors: Babak Maboudi Afkham and Julianne Chung and Matthias Chung
- Abstract summary: We consider a supervised learning approach, where a network is trained to approximate the mapping from observation data to regularization parameters.
We show that a wide variety of regularization functionals, forward models, and noise models may be considered.
The network-obtained regularization parameters can be computed more efficiently and may even lead to more accurate solutions.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work, we describe a new approach that uses deep neural networks (DNN)
to obtain regularization parameters for solving inverse problems. We consider a
supervised learning approach, where a network is trained to approximate the
mapping from observation data to regularization parameters. Once the network is
trained, regularization parameters for newly obtained data can be computed by
efficient forward propagation of the DNN. We show that a wide variety of
regularization functionals, forward models, and noise models may be considered.
The network-obtained regularization parameters can be computed more efficiently
and may even lead to more accurate solutions compared to existing
regularization parameter selection methods. We emphasize that the key advantage
of using DNNs for learning regularization parameters, compared to previous
works on learning via optimal experimental design or empirical Bayes risk
minimization, is greater generalizability. That is, rather than computing one
set of parameters that is optimal with respect to one particular design
objective, DNN-computed regularization parameters are tailored to the specific
features or properties of the newly observed data. Thus, our approach may
better handle cases where the observation is not a close representation of the
training set. Furthermore, we avoid the need for expensive and challenging
bilevel optimization methods as utilized in other existing training approaches.
Numerical results demonstrate the potential of using DNNs to learn
regularization parameters.
Related papers
- Sparse Deep Learning Models with the $\ell_1$ Regularization [6.268040192346312]
Sparse neural networks are highly desirable in deep learning.
We study how choices of regularization parameters influence the sparsity level of learned neural networks.
arXiv Detail & Related papers (2024-08-05T19:38:45Z) - Neural Parameter Regression for Explicit Representations of PDE Solution Operators [22.355460388065964]
We introduce Neural Regression (NPR), a novel framework specifically developed for learning solution operators in Partial Differential Equations (PDEs)
NPR employs Physics-Informed Neural Network (PINN, Raissi et al., 2021) techniques to regress Neural Network (NN) parameters.
The framework shows remarkable adaptability to new initial and boundary conditions, allowing for rapid fine-tuning and inference.
arXiv Detail & Related papers (2024-03-19T14:30:56Z) - Continual Learning via Sequential Function-Space Variational Inference [65.96686740015902]
We propose an objective derived by formulating continual learning as sequential function-space variational inference.
Compared to objectives that directly regularize neural network predictions, the proposed objective allows for more flexible variational distributions.
We demonstrate that, across a range of task sequences, neural networks trained via sequential function-space variational inference achieve better predictive accuracy than networks trained with related methods.
arXiv Detail & Related papers (2023-12-28T18:44:32Z) - Function-Space Regularization in Neural Networks: A Probabilistic
Perspective [51.133793272222874]
We show that we can derive a well-motivated regularization technique that allows explicitly encoding information about desired predictive functions into neural network training.
We evaluate the utility of this regularization technique empirically and demonstrate that the proposed method leads to near-perfect semantic shift detection and highly-calibrated predictive uncertainty estimates.
arXiv Detail & Related papers (2023-12-28T17:50:56Z) - Learning Low Dimensional State Spaces with Overparameterized Recurrent
Neural Nets [57.06026574261203]
We provide theoretical evidence for learning low-dimensional state spaces, which can also model long-term memory.
Experiments corroborate our theory, demonstrating extrapolation via learning low-dimensional state spaces with both linear and non-linear RNNs.
arXiv Detail & Related papers (2022-10-25T14:45:15Z) - Learning to Learn with Generative Models of Neural Network Checkpoints [71.06722933442956]
We construct a dataset of neural network checkpoints and train a generative model on the parameters.
We find that our approach successfully generates parameters for a wide range of loss prompts.
We apply our method to different neural network architectures and tasks in supervised and reinforcement learning.
arXiv Detail & Related papers (2022-09-26T17:59:58Z) - Sparse Bayesian Deep Learning for Dynamic System Identification [14.040914364617418]
This paper proposes a sparse Bayesian treatment of deep neural networks (DNNs) for system identification.
The proposed Bayesian approach offers a principled way to alleviate the challenges by marginal likelihood/model evidence approximation.
The effectiveness of the proposed Bayesian approach is demonstrated on several linear and nonlinear systems identification benchmarks.
arXiv Detail & Related papers (2021-07-27T16:09:48Z) - Online hyperparameter optimization by real-time recurrent learning [57.01871583756586]
Our framework takes advantage of the analogy between hyperparameter optimization and parameter learning in neural networks (RNNs)
It adapts a well-studied family of online learning algorithms for RNNs to tune hyperparameters and network parameters simultaneously.
This procedure yields systematically better generalization performance compared to standard methods, at a fraction of wallclock time.
arXiv Detail & Related papers (2021-02-15T19:36:18Z) - Provably Efficient Neural Estimation of Structural Equation Model: An
Adversarial Approach [144.21892195917758]
We study estimation in a class of generalized Structural equation models (SEMs)
We formulate the linear operator equation as a min-max game, where both players are parameterized by neural networks (NNs), and learn the parameters of these neural networks using a gradient descent.
For the first time we provide a tractable estimation procedure for SEMs based on NNs with provable convergence and without the need for sample splitting.
arXiv Detail & Related papers (2020-07-02T17:55:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.