A data-driven choice of misfit function for FWI using reinforcement
learning
- URL: http://arxiv.org/abs/2002.03154v1
- Date: Sat, 8 Feb 2020 12:31:33 GMT
- Title: A data-driven choice of misfit function for FWI using reinforcement
learning
- Authors: Bingbing Sun and Tariq Alkhalifah
- Abstract summary: We use a deep-Q network (DQN) to learn an optimal policy to determine the proper timing to switch between different misfit functions.
Specifically, we train the state-action value function (Q) to predict when to use the conventional L2-norm misfit function or the more advanced optimal-transport matching-filter (OTMF) misfit.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the workflow of Full-Waveform Inversion (FWI), we often tune the
parameters of the inversion to help us avoid cycle skipping and obtain high
resolution models. For example, typically start by using objective functions
that avoid cycle skipping, like tomographic and image based or using only low
frequency, and then later, we utilize the least squares misfit to admit high
resolution information. We also may perform an isotropic (acoustic) inversion
to first update the velocity model and then switch to multi-parameter
anisotropic (elastic) inversions to fully recover the complex physics. Such
hierarchical approaches are common in FWI, and they often depend on our manual
intervention based on many factors, and of course, results depend on
experience. However, with the large data size often involved in the inversion
and the complexity of the process, making optimal choices is difficult even for
an experienced practitioner. Thus, as an example, and within the framework of
reinforcement learning, we utilize a deep-Q network (DQN) to learn an optimal
policy to determine the proper timing to switch between different misfit
functions. Specifically, we train the state-action value function (Q) to
predict when to use the conventional L2-norm misfit function or the more
advanced optimal-transport matching-filter (OTMF) misfit to mitigate the
cycle-skipping and obtain high resolution, as well as improve convergence. We
use a simple while demonstrative shifted-signal inversion examples to
demonstrate the basic principles of the proposed method.
Related papers
- LeRF: Learning Resampling Function for Adaptive and Efficient Image Interpolation [64.34935748707673]
Recent deep neural networks (DNNs) have made impressive progress in performance by introducing learned data priors.
We propose a novel method of Learning Resampling (termed LeRF) which takes advantage of both the structural priors learned by DNNs and the locally continuous assumption.
LeRF assigns spatially varying resampling functions to input image pixels and learns to predict the shapes of these resampling functions with a neural network.
arXiv Detail & Related papers (2024-07-13T16:09:45Z) - One-Dimensional Deep Image Prior for Curve Fitting of S-Parameters from
Electromagnetic Solvers [57.441926088870325]
Deep Image Prior (DIP) is a technique that optimized the weights of a randomly-d convolutional neural network to fit a signal from noisy or under-determined measurements.
Relative to publicly available implementations of Vector Fitting (VF), our method shows superior performance on nearly all test examples.
arXiv Detail & Related papers (2023-06-06T20:28:37Z) - End-to-End Meta-Bayesian Optimisation with Transformer Neural Processes [52.818579746354665]
This paper proposes the first end-to-end differentiable meta-BO framework that generalises neural processes to learn acquisition functions via transformer architectures.
We enable this end-to-end framework with reinforcement learning (RL) to tackle the lack of labelled acquisition data.
arXiv Detail & Related papers (2023-05-25T10:58:46Z) - NL-CS Net: Deep Learning with Non-Local Prior for Image Compressive
Sensing [7.600617428107161]
Deep learning has been applied to compressive sensing (CS) of images successfully in recent years.
This paper proposes a novel CS method using non-local prior which combines the interpretability of the traditional optimization methods with the speed of network-based methods, called NL-CS Net.
arXiv Detail & Related papers (2023-05-06T02:34:28Z) - Faster Adaptive Federated Learning [84.38913517122619]
Federated learning has attracted increasing attention with the emergence of distributed data.
In this paper, we propose an efficient adaptive algorithm (i.e., FAFED) based on momentum-based variance reduced technique in cross-silo FL.
arXiv Detail & Related papers (2022-12-02T05:07:50Z) - Incremental Spatial and Spectral Learning of Neural Operators for
Solving Large-Scale PDEs [86.35471039808023]
We introduce the Incremental Fourier Neural Operator (iFNO), which progressively increases the number of frequency modes used by the model.
We show that iFNO reduces total training time while maintaining or improving generalization performance across various datasets.
Our method demonstrates a 10% lower testing error, using 20% fewer frequency modes compared to the existing Fourier Neural Operator, while also achieving a 30% faster training.
arXiv Detail & Related papers (2022-11-28T09:57:15Z) - Transform Once: Efficient Operator Learning in Frequency Domain [69.74509540521397]
We study deep neural networks designed to harness the structure in frequency domain for efficient learning of long-range correlations in space or time.
This work introduces a blueprint for frequency domain learning through a single transform: transform once (T1)
arXiv Detail & Related papers (2022-11-26T01:56:05Z) - Loop Unrolled Shallow Equilibrium Regularizer (LUSER) -- A
Memory-Efficient Inverse Problem Solver [26.87738024952936]
In inverse problems we aim to reconstruct some underlying signal of interest from potentially corrupted and often ill-posed measurements.
We propose an LU algorithm with shallow equilibrium regularizers (L)
These implicit models are as expressive as deeper convolutional networks, but far more memory efficient during training.
arXiv Detail & Related papers (2022-10-10T19:50:37Z) - Obtaining Adjustable Regularization for Free via Iterate Averaging [43.75491612671571]
Regularization for optimization is a crucial technique to avoid overfitting in machine learning.
We establish an averaging scheme that converts the iterates of SGD on an arbitrary strongly convex and smooth objective function to its regularized counterpart.
Our approaches can be used for accelerated and preconditioned optimization methods as well.
arXiv Detail & Related papers (2020-08-15T15:28:05Z) - ML-misfit: Learning a robust misfit function for full-waveform inversion
using machine learning [0.0]
We learn a misfit function for full waveform inversion (FWI) based on machine learning.
Inspired by the optimal transport of the matching filter misfit, we design a neural network (NN) architecture for the misfit function.
We demonstrate the effectiveness and robustness of the learned ML-misfit by applying it to the well-known Marmousi model.
arXiv Detail & Related papers (2020-02-08T13:27:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.