Data-driven super-parameterization using deep learning: Experimentation
with multi-scale Lorenz 96 systems and transfer-learning
- URL: http://arxiv.org/abs/2002.11167v1
- Date: Tue, 25 Feb 2020 20:43:42 GMT
- Title: Data-driven super-parameterization using deep learning: Experimentation
with multi-scale Lorenz 96 systems and transfer-learning
- Authors: Ashesh Chattopadhyay, Adam Subel, Pedram Hassanzadeh
- Abstract summary: We propose a data-driven SP (DD-SP) to make weather/climate modeling computationally affordable.
With the same computational cost, DD-SP substantially outperforms LR, and is better than DD-P, particularly when scale separation is lacking.
DD-SP is much cheaper than SP, yet its accuracy is the same in reproducing long-term statistics and often comparable in short-term forecasting.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: To make weather/climate modeling computationally affordable, small-scale
processes are usually represented in terms of the large-scale,
explicitly-resolved processes using physics-based or semi-empirical
parameterization schemes. Another approach, computationally more demanding but
often more accurate, is super-parameterization (SP), which involves integrating
the equations of small-scale processes on high-resolution grids embedded within
the low-resolution grids of large-scale processes. Recently, studies have used
machine learning (ML) to develop data-driven parameterization (DD-P) schemes.
Here, we propose a new approach, data-driven SP (DD-SP), in which the equations
of the small-scale processes are integrated data-drivenly using ML methods such
as recurrent neural networks. Employing multi-scale Lorenz 96 systems as
testbed, we compare the cost and accuracy (in terms of both short-term
prediction and long-term statistics) of parameterized low-resolution (LR), SP,
DD-P, and DD-SP models. We show that with the same computational cost, DD-SP
substantially outperforms LR, and is better than DD-P, particularly when scale
separation is lacking. DD-SP is much cheaper than SP, yet its accuracy is the
same in reproducing long-term statistics and often comparable in short-term
forecasting. We also investigate generalization, finding that when models
trained on data from one system are applied to a system with different forcing
(e.g., more chaotic), the models often do not generalize, particularly when the
short-term prediction accuracy is examined. But we show that transfer-learning,
which involves re-training the data-driven model with a small amount of data
from the new system, significantly improves generalization. Potential
applications of DD-SP and transfer-learning in climate/weather modeling and the
expected challenges are discussed.
Related papers
- A Bayesian Approach to Data Point Selection [24.98069363998565]
Data point selection (DPS) is becoming a critical topic in deep learning.
Existing approaches to DPS are predominantly based on a bi-level optimisation (BLO) formulation.
We propose a novel Bayesian approach to DPS.
arXiv Detail & Related papers (2024-11-06T09:04:13Z) - Computation-Aware Gaussian Processes: Model Selection And Linear-Time Inference [55.150117654242706]
We show that model selection for computation-aware GPs trained on 1.8 million data points can be done within a few hours on a single GPU.
As a result of this work, Gaussian processes can be trained on large-scale datasets without significantly compromising their ability to quantify uncertainty.
arXiv Detail & Related papers (2024-11-01T21:11:48Z) - Parallel and Limited Data Voice Conversion Using Stochastic Variational
Deep Kernel Learning [2.5782420501870296]
This paper proposes a voice conversion method that works with limited data.
It is based on variational deep kernel learning (SVDKL)
It is possible to estimate non-smooth and more complex functions.
arXiv Detail & Related papers (2023-09-08T16:32:47Z) - FaDIn: Fast Discretized Inference for Hawkes Processes with General
Parametric Kernels [82.53569355337586]
This work offers an efficient solution to temporal point processes inference using general parametric kernels with finite support.
The method's effectiveness is evaluated by modeling the occurrence of stimuli-induced patterns from brain signals recorded with magnetoencephalography (MEG)
Results show that the proposed approach leads to an improved estimation of pattern latency than the state-of-the-art.
arXiv Detail & Related papers (2022-10-10T12:35:02Z) - Neural parameter calibration for large-scale multi-agent models [0.7734726150561089]
We present a method to retrieve accurate probability densities for parameters using neural equations.
The two combined create a powerful tool that can quickly estimate densities on model parameters, even for very large systems.
arXiv Detail & Related papers (2022-09-27T17:36:26Z) - Semi-Parametric Inducing Point Networks and Neural Processes [15.948270454686197]
Semi-parametric inducing point networks (SPIN) can query the training set at inference time in a compute-efficient manner.
SPIN attains linear complexity via a cross-attention mechanism between datapoints inspired by inducing point methods.
In our experiments, SPIN reduces memory requirements, improves accuracy across a range of meta-learning tasks, and improves state-of-the-art performance on an important practical problem, genotype imputation.
arXiv Detail & Related papers (2022-05-24T01:42:46Z) - Scaling Structured Inference with Randomization [64.18063627155128]
We propose a family of dynamic programming (RDP) randomized for scaling structured models to tens of thousands of latent states.
Our method is widely applicable to classical DP-based inference.
It is also compatible with automatic differentiation so can be integrated with neural networks seamlessly.
arXiv Detail & Related papers (2021-12-07T11:26:41Z) - Large-scale Neural Solvers for Partial Differential Equations [48.7576911714538]
Solving partial differential equations (PDE) is an indispensable part of many branches of science as many processes can be modelled in terms of PDEs.
Recent numerical solvers require manual discretization of the underlying equation as well as sophisticated, tailored code for distributed computing.
We examine the applicability of continuous, mesh-free neural solvers for partial differential equations, physics-informed neural networks (PINNs)
We discuss the accuracy of GatedPINN with respect to analytical solutions -- as well as state-of-the-art numerical solvers, such as spectral solvers.
arXiv Detail & Related papers (2020-09-08T13:26:51Z) - Extrapolation for Large-batch Training in Deep Learning [72.61259487233214]
We show that a host of variations can be covered in a unified framework that we propose.
We prove the convergence of this novel scheme and rigorously evaluate its empirical performance on ResNet, LSTM, and Transformer.
arXiv Detail & Related papers (2020-06-10T08:22:41Z) - Transformer Hawkes Process [79.16290557505211]
We propose a Transformer Hawkes Process (THP) model, which leverages the self-attention mechanism to capture long-term dependencies.
THP outperforms existing models in terms of both likelihood and event prediction accuracy by a notable margin.
We provide a concrete example, where THP achieves improved prediction performance for learning multiple point processes when incorporating their relational information.
arXiv Detail & Related papers (2020-02-21T13:48:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.