Implicit Full Waveform Inversion with Deep Neural Representation
- URL: http://arxiv.org/abs/2209.03525v1
- Date: Thu, 8 Sep 2022 01:54:50 GMT
- Title: Implicit Full Waveform Inversion with Deep Neural Representation
- Authors: Jian Sun and Kristopher Innanen
- Abstract summary: We propose the implicit full waveform inversion (IFWI) algorithm using continuously and implicitly defined deep neural representations.
Both theoretical and experimental analyses indicates that, given a random initial model, IFWI is able to converge to the global minimum.
IFWI has a certain degree of robustness and strong generalization ability that are exemplified in the experiments of various 2D geological models.
- Score: 91.3755431537592
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Full waveform inversion (FWI) commonly stands for the state-of-the-art
approach for imaging subsurface structures and physical parameters, however,
its implementation usually faces great challenges, such as building a good
initial model to escape from local minima, and evaluating the uncertainty of
inversion results. In this paper, we propose the implicit full waveform
inversion (IFWI) algorithm using continuously and implicitly defined deep
neural representations. Compared to FWI, which is sensitive to the initial
model, IFWI benefits from the increased degrees of freedom with deep learning
optimization, thus allowing to start from a random initialization, which
greatly reduces the risk of non-uniqueness and being trapped in local minima.
Both theoretical and experimental analyses indicates that, given a random
initial model, IFWI is able to converge to the global minimum and produce a
high-resolution image of subsurface with fine structures. In addition,
uncertainty analysis of IFWI can be easily performed by approximating Bayesian
inference with various deep learning approaches, which is analyzed in this
paper by adding dropout neurons. Furthermore, IFWI has a certain degree of
robustness and strong generalization ability that are exemplified in the
experiments of various 2D geological models. With proper setup, IFWI can also
be well suited for multi-scale joint geophysical inversion.
Related papers
- PreAdaptFWI: Pretrained-Based Adaptive Residual Learning for Full-Waveform Inversion Without Dataset Dependency [8.719356558714246]
Full-waveform inversion (FWI) is a method that utilizes seismic data to invert the physical parameters of subsurface media.
Due to its ill-posed nature, FWI is susceptible to getting trapped in local minima.
Various research efforts have attempted to combine neural networks with FWI to stabilize the inversion process.
arXiv Detail & Related papers (2025-02-17T15:30:17Z) - Geometric Neural Process Fields [58.77241763774756]
Geometric Neural Process Fields (G-NPF) is a probabilistic framework for neural radiance fields that explicitly captures uncertainty.
Building on these bases, we design a hierarchical latent variable model, allowing G-NPF to integrate structural information across multiple spatial levels.
Experiments on novel-view synthesis for 3D scenes, as well as 2D image and 1D signal regression, demonstrate the effectiveness of our method.
arXiv Detail & Related papers (2025-02-04T14:17:18Z) - Automatic Differentiation-based Full Waveform Inversion with Flexible Workflows [6.41869310373004]
Full waveform inversion (FWI) is able to construct high-resolution subsurface models by minimizing discrepancies between observed and simulated seismic data.
Recently, automatic differentiation (AD) has proven to be effective in simplifying solutions of various inverse problems, including FWI.
We present an open-source AD-based FWI framework (ADFWI), which is designed to simplify the design, development, and evaluation of novel approaches in FWI with flexibility.
arXiv Detail & Related papers (2024-11-30T13:58:41Z) - Structured Optimal Variational Inference for Dynamic Latent Space Models [16.531262817315696]
We consider a latent space model for dynamic networks, where our objective is to estimate the pairwise inner products plus the intercept of the latent positions.
To balance posterior inference and computational scalability, we consider a structured mean-field variational inference framework.
arXiv Detail & Related papers (2022-09-29T22:10:42Z) - Robust lEarned Shrinkage-Thresholding (REST): Robust unrolling for
sparse recover [87.28082715343896]
We consider deep neural networks for solving inverse problems that are robust to forward model mis-specifications.
We design a new robust deep neural network architecture by applying algorithm unfolding techniques to a robust version of the underlying recovery problem.
The proposed REST network is shown to outperform state-of-the-art model-based and data-driven algorithms in both compressive sensing and radar imaging problems.
arXiv Detail & Related papers (2021-10-20T06:15:45Z) - Revisit Geophysical Imaging in A New View of Physics-informed Generative
Adversarial Learning [2.12121796606941]
Full waveform inversion produces high-resolution subsurface models.
FWI with least-squares function suffers from many drawbacks such as the local-minima problem.
Recent works relying on partial differential equations and neural networks show promising performance for two-dimensional FWI.
We propose an unsupervised learning paradigm that integrates wave equation with a discriminate network to accurately estimate the physically consistent models.
arXiv Detail & Related papers (2021-09-23T15:54:40Z) - Influence Estimation and Maximization via Neural Mean-Field Dynamics [60.91291234832546]
We propose a novel learning framework using neural mean-field (NMF) dynamics for inference and estimation problems.
Our framework can simultaneously learn the structure of the diffusion network and the evolution of node infection probabilities.
arXiv Detail & Related papers (2021-06-03T00:02:05Z) - Generalization bound of globally optimal non-convex neural network
training: Transportation map estimation by infinite dimensional Langevin
dynamics [50.83356836818667]
We introduce a new theoretical framework to analyze deep learning optimization with connection to its generalization error.
Existing frameworks such as mean field theory and neural tangent kernel theory for neural network optimization analysis typically require taking limit of infinite width of the network to show its global convergence.
arXiv Detail & Related papers (2020-07-11T18:19:50Z) - DessiLBI: Exploring Structural Sparsity of Deep Networks via
Differential Inclusion Paths [45.947140164621096]
We propose a new approach based on differential inclusions of inverse scale spaces.
We show that DessiLBI unveils "winning tickets" in early epochs.
arXiv Detail & Related papers (2020-07-04T04:40:16Z) - Modeling from Features: a Mean-field Framework for Over-parameterized
Deep Neural Networks [54.27962244835622]
This paper proposes a new mean-field framework for over- parameterized deep neural networks (DNNs)
In this framework, a DNN is represented by probability measures and functions over its features in the continuous limit.
We illustrate the framework via the standard DNN and the Residual Network (Res-Net) architectures.
arXiv Detail & Related papers (2020-07-03T01:37:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.