Adjoint-Matching Neural Network Surrogates for Fast 4D-Var Data
Assimilation
- URL: http://arxiv.org/abs/2111.08626v1
- Date: Tue, 16 Nov 2021 17:11:05 GMT
- Title: Adjoint-Matching Neural Network Surrogates for Fast 4D-Var Data
Assimilation
- Authors: Austin Chennault, Andrey A. Popov, Amit N. Subrahmanya, Rachel Cooper,
Anuj Karpatne, Adrian Sandu
- Abstract summary: We formulate and analyze approaches incorporating derivative information into the construction of neural network surrogates.
Two methods demonstrate superior performance when compared with a surrogate network trained without adjoint information.
- Score: 1.7416597120949546
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The data assimilation procedures used in many operational numerical weather
forecasting systems are based around variants of the 4D-Var algorithm. The cost
of solving the 4D-Var problem is dominated by the cost of forward and adjoint
evaluations of the physical model. This motivates their substitution by fast,
approximate surrogate models. Neural networks offer a promising approach for
the data-driven creation of surrogate models. The accuracy of the surrogate
4D-Var problem's solution has been shown to depend explicitly on accurate
modeling of the forward and adjoint for other surrogate modeling approaches and
in the general nonlinear setting. We formulate and analyze several approaches
to incorporating derivative information into the construction of neural network
surrogates. The resulting networks are tested on out of training set data and
in a sequential data assimilation setting on the Lorenz-63 system. Two methods
demonstrate superior performance when compared with a surrogate network trained
without adjoint information, showing the benefit of incorporating adjoint
information into the training process.
Related papers
- Neural Network Reprogrammability: A Unified Theme on Model Reprogramming, Prompt Tuning, and Prompt Instruction [55.914891182214475]
We introduce neural network reprogrammability as a unifying framework for model adaptation.<n>We present a taxonomy that categorizes such information manipulation approaches across four key dimensions.<n>We also analyze remaining technical challenges and ethical considerations.
arXiv Detail & Related papers (2025-06-05T05:42:27Z) - Tensor-Var: Efficient Four-Dimensional Variational Data Assimilation [30.63086465547801]
Four-dimensional variational assimilation (4D-Var) faces high computational costs in complex nonlinear systems and depends on imperfect state-observation mappings.<n>Deep learning (DL) offers more expressive approximators, while integrating DL models into 4D-Var is challenging due to their nonlinearities and lack of theoretical guarantees in assimilation results.<n>We propose a novel framework that integrates kernel conditional mean embedding (CME) with 4D-Var to linearize nonlinear dynamics, achieving convex optimization in a learned feature space.
arXiv Detail & Related papers (2025-01-23T01:43:31Z) - Diffusion models as probabilistic neural operators for recovering unobserved states of dynamical systems [49.2319247825857]
We show that diffusion-based generative models exhibit many properties favourable for neural operators.
We propose to train a single model adaptable to multiple tasks, by alternating between the tasks during training.
arXiv Detail & Related papers (2024-05-11T21:23:55Z) - Solving Inverse Problems with Model Mismatch using Untrained Neural Networks within Model-based Architectures [14.551812310439004]
We introduce an untrained forward model residual block within the model-based architecture to match the data consistency in the measurement domain for each instance.
Our approach offers a unified solution that is less parameter-sensitive, requires no additional data, and enables simultaneous fitting of the forward model and reconstruction in a single pass.
arXiv Detail & Related papers (2024-03-07T19:02:13Z) - Diffusion-Model-Assisted Supervised Learning of Generative Models for
Density Estimation [10.793646707711442]
We present a framework for training generative models for density estimation.
We use the score-based diffusion model to generate labeled data.
Once the labeled data are generated, we can train a simple fully connected neural network to learn the generative model in the supervised manner.
arXiv Detail & Related papers (2023-10-22T23:56:19Z) - Training Deep Surrogate Models with Large Scale Online Learning [48.7576911714538]
Deep learning algorithms have emerged as a viable alternative for obtaining fast solutions for PDEs.
Models are usually trained on synthetic data generated by solvers, stored on disk and read back for training.
It proposes an open source online training framework for deep surrogate models.
arXiv Detail & Related papers (2023-06-28T12:02:27Z) - VTAE: Variational Transformer Autoencoder with Manifolds Learning [144.0546653941249]
Deep generative models have demonstrated successful applications in learning non-linear data distributions through a number of latent variables.
The nonlinearity of the generator implies that the latent space shows an unsatisfactory projection of the data space, which results in poor representation learning.
We show that geodesics and accurate computation can substantially improve the performance of deep generative models.
arXiv Detail & Related papers (2023-04-03T13:13:19Z) - Online model error correction with neural networks in the incremental
4D-Var framework [0.0]
We develop a new weak-constraint 4D-Var formulation which can be used to train a neural network for online model error correction.
The method is implemented in the ECMWF Object-Oriented Prediction System.
The results confirm that online learning is effective and yields a more accurate model error correction than offline learning.
arXiv Detail & Related papers (2022-10-25T07:45:33Z) - Learning to Learn with Generative Models of Neural Network Checkpoints [71.06722933442956]
We construct a dataset of neural network checkpoints and train a generative model on the parameters.
We find that our approach successfully generates parameters for a wide range of loss prompts.
We apply our method to different neural network architectures and tasks in supervised and reinforcement learning.
arXiv Detail & Related papers (2022-09-26T17:59:58Z) - Leveraging Intrinsic Gradient Information for Machine Learning Model
Training [4.682734815593623]
derivatives of target variables with respect to inputs can be leveraged to improve the accuracy of differentiable machine learning models.
Four key ideas are explored: (1) Improving the predictive accuracy of linear regression models and feed-forward neural networks (NNs); (2) Using the difference between the performance of feedforward NNs trained with and without gradient information to tune NN complexity; and (4) Using gradient information to improve generative image models.
arXiv Detail & Related papers (2021-11-30T20:50:45Z) - Model Fusion with Kullback--Leibler Divergence [58.20269014662046]
We propose a method to fuse posterior distributions learned from heterogeneous datasets.
Our algorithm relies on a mean field assumption for both the fused model and the individual dataset posteriors.
arXiv Detail & Related papers (2020-07-13T03:27:45Z) - Pre-Trained Models for Heterogeneous Information Networks [57.78194356302626]
We propose a self-supervised pre-training and fine-tuning framework, PF-HIN, to capture the features of a heterogeneous information network.
PF-HIN consistently and significantly outperforms state-of-the-art alternatives on each of these tasks, on four datasets.
arXiv Detail & Related papers (2020-07-07T03:36:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.