ML-misfit: Learning a robust misfit function for full-waveform inversion
using machine learning
- URL: http://arxiv.org/abs/2002.03163v2
- Date: Wed, 18 Mar 2020 10:53:05 GMT
- Title: ML-misfit: Learning a robust misfit function for full-waveform inversion
using machine learning
- Authors: Bingbing Sun and Tariq Alkhalifah
- Abstract summary: We learn a misfit function for full waveform inversion (FWI) based on machine learning.
Inspired by the optimal transport of the matching filter misfit, we design a neural network (NN) architecture for the misfit function.
We demonstrate the effectiveness and robustness of the learned ML-misfit by applying it to the well-known Marmousi model.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Most of the available advanced misfit functions for full waveform inversion
(FWI) are hand-crafted, and the performance of those misfit functions is
data-dependent. Thus, we propose to learn a misfit function for FWI, entitled
ML-misfit, based on machine learning. Inspired by the optimal transport of the
matching filter misfit, we design a neural network (NN) architecture for the
misfit function in a form similar to comparing the mean and variance for two
distributions. To guarantee the resulting learned misfit is a metric, we
accommodate the symmetry of the misfit with respect to its input and a Hinge
loss regularization term in a meta-loss function to satisfy the "triangle
inequality" rule. In the framework of meta-learning, we train the network by
running FWI to invert for randomly generated velocity models and update the
parameters of the NN by minimizing the meta-loss, which is defined as
accumulated difference between the true and inverted models. We first
illustrate the basic principle of the ML-misfit for learning a convex misfit
function for travel-time shifted signals. Further, we train the NN on 2D
horizontally layered models, and we demonstrate the effectiveness and
robustness of the learned ML-misfit by applying it to the well-known Marmousi
model.
Related papers
- A Stochastic Approach to Bi-Level Optimization for Hyperparameter Optimization and Meta Learning [74.80956524812714]
We tackle the general differentiable meta learning problem that is ubiquitous in modern deep learning.
These problems are often formalized as Bi-Level optimizations (BLO)
We introduce a novel perspective by turning a given BLO problem into a ii optimization, where the inner loss function becomes a smooth distribution, and the outer loss becomes an expected loss over the inner distribution.
arXiv Detail & Related papers (2024-10-14T12:10:06Z) - Bring Metric Functions into Diffusion Models [145.71911023514252]
We introduce a Cascaded Diffusion Model (Cas-DM) that improves a Denoising Diffusion Probabilistic Model (DDPM)
The proposed diffusion model backbone enables the effective use of the LPIPS loss, leading to state-of-the-art image quality (FID, sFID, IS)
Experiment results show that the proposed diffusion model backbone enables the effective use of the LPIPS loss, leading to state-of-the-art image quality (FID, sFID, IS)
arXiv Detail & Related papers (2024-01-04T18:55:01Z) - Learning Layer-wise Equivariances Automatically using Gradients [66.81218780702125]
Convolutions encode equivariance symmetries into neural networks leading to better generalisation performance.
symmetries provide fixed hard constraints on the functions a network can represent, need to be specified in advance, and can not be adapted.
Our goal is to allow flexible symmetry constraints that can automatically be learned from data using gradients.
arXiv Detail & Related papers (2023-10-09T20:22:43Z) - Theoretical Characterization of the Generalization Performance of
Overfitted Meta-Learning [70.52689048213398]
This paper studies the performance of overfitted meta-learning under a linear regression model with Gaussian features.
We find new and interesting properties that do not exist in single-task linear regression.
Our analysis suggests that benign overfitting is more significant and easier to observe when the noise and the diversity/fluctuation of the ground truth of each training task are large.
arXiv Detail & Related papers (2023-04-09T20:36:13Z) - An Adaptive Plug-and-Play Network for Few-Shot Learning [12.023266104119289]
Few-shot learning requires a model to classify new samples after learning from only a few samples.
Deep networks and complex metrics tend to induce overfitting, making it difficult to further improve the performance.
We propose plug-and-play model-adaptive resizer (MAR) and adaptive similarity metric (ASM) without any other losses.
arXiv Detail & Related papers (2023-02-18T13:25:04Z) - Evolving parametrized Loss for Image Classification Learning on Small
Datasets [1.4685355149711303]
This paper proposes a meta-learning approach to evolving a parametrized loss function, which is called Meta-Loss Network (MLN)
In our approach, the MLN is embedded in the framework of classification learning as a differentiable objective function.
Experiment results demonstrate that the MLN effectively improved generalization compared to classical cross-entropy error and mean squared error.
arXiv Detail & Related papers (2021-03-15T10:00:18Z) - B-SMALL: A Bayesian Neural Network approach to Sparse Model-Agnostic
Meta-Learning [2.9189409618561966]
We propose a Bayesian neural network based MAML algorithm, which we refer to as the B-SMALL algorithm.
We demonstrate the performance of B-MAML using classification and regression tasks, and highlight that training a sparsifying BNN using MAML indeed improves the parameter footprint of the model.
arXiv Detail & Related papers (2021-01-01T09:19:48Z) - Shaping Deep Feature Space towards Gaussian Mixture for Visual
Classification [74.48695037007306]
We propose a Gaussian mixture (GM) loss function for deep neural networks for visual classification.
With a classification margin and a likelihood regularization, the GM loss facilitates both high classification performance and accurate modeling of the feature distribution.
The proposed model can be implemented easily and efficiently without using extra trainable parameters.
arXiv Detail & Related papers (2020-11-18T03:32:27Z) - Optimization-driven Machine Learning for Intelligent Reflecting Surfaces
Assisted Wireless Networks [82.33619654835348]
Intelligent surface (IRS) has been employed to reshape the wireless channels by controlling individual scattering elements' phase shifts.
Due to the large size of scattering elements, the passive beamforming is typically challenged by the high computational complexity.
In this article, we focus on machine learning (ML) approaches for performance in IRS-assisted wireless networks.
arXiv Detail & Related papers (2020-08-29T08:39:43Z) - A data-driven choice of misfit function for FWI using reinforcement
learning [0.0]
We use a deep-Q network (DQN) to learn an optimal policy to determine the proper timing to switch between different misfit functions.
Specifically, we train the state-action value function (Q) to predict when to use the conventional L2-norm misfit function or the more advanced optimal-transport matching-filter (OTMF) misfit.
arXiv Detail & Related papers (2020-02-08T12:31:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.