Non-parametric Kernel-Based Estimation of Probability Distributions for
Precipitation Modeling
- URL: http://arxiv.org/abs/2109.09961v1
- Date: Tue, 21 Sep 2021 04:52:00 GMT
- Title: Non-parametric Kernel-Based Estimation of Probability Distributions for
Precipitation Modeling
- Authors: Andrew Pavlides, Vasiliki Agou, Dionissios T. Hristopulos
- Abstract summary: We derive non-parametric estimates of the cumulative distribution function (CDF) of precipitation amount for wet time intervals.
We show that KCDE provides better estimates of the probability distribution than the standard empirical (staircase) estimate.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The probability distribution of precipitation amount strongly depends on
geography, climate zone, and time scale considered. Closed-form parametric
probability distributions are not sufficiently flexible to provide accurate and
universal models for precipitation amount over different time scales. In this
paper we derive non-parametric estimates of the cumulative distribution
function (CDF) of precipitation amount for wet time intervals. The CDF
estimates are obtained by integrating the kernel density estimator leading to
semi-explicit CDF expressions for different kernel functions. We investigate
kernel-based CDF estimation with an adaptive plug-in bandwidth (KCDE), using
both synthetic data sets and reanalysis precipitation data from the island of
Crete (Greece). We show that KCDE provides better estimates of the probability
distribution than the standard empirical (staircase) estimate and kernel-based
estimates that use the normal reference bandwidth. We also demonstrate that
KCDE enables the simulation of non-parametric precipitation amount
distributions by means of the inverse transform sampling method.
Related papers
- Theory on Score-Mismatched Diffusion Models and Zero-Shot Conditional Samplers [49.97755400231656]
We present the first performance guarantee with explicit dimensional general score-mismatched diffusion samplers.
We show that score mismatches result in an distributional bias between the target and sampling distributions, proportional to the accumulated mismatch between the target and training distributions.
This result can be directly applied to zero-shot conditional samplers for any conditional model, irrespective of measurement noise.
arXiv Detail & Related papers (2024-10-17T16:42:12Z) - Convergence of Score-Based Discrete Diffusion Models: A Discrete-Time Analysis [56.442307356162864]
We study the theoretical aspects of score-based discrete diffusion models under the Continuous Time Markov Chain (CTMC) framework.
We introduce a discrete-time sampling algorithm in the general state space $[S]d$ that utilizes score estimators at predefined time points.
Our convergence analysis employs a Girsanov-based method and establishes key properties of the discrete score function.
arXiv Detail & Related papers (2024-10-03T09:07:13Z) - A Likelihood Based Approach to Distribution Regression Using Conditional Deep Generative Models [6.647819824559201]
We study the large-sample properties of a likelihood-based approach for estimating conditional deep generative models.
Our results lead to the convergence rate of a sieve maximum likelihood estimator for estimating the conditional distribution.
arXiv Detail & Related papers (2024-10-02T20:46:21Z) - Probabilistic Forecasting with Stochastic Interpolants and Föllmer Processes [18.344934424278048]
We propose a framework for probabilistic forecasting of dynamical systems based on generative modeling.
We show that the drift and the diffusion coefficients of this SDE can be adjusted after training, and that a specific choice that minimizes the impact of the estimation error gives a F"ollmer process.
arXiv Detail & Related papers (2024-03-20T16:33:06Z) - Score Approximation, Estimation and Distribution Recovery of Diffusion
Models on Low-Dimensional Data [68.62134204367668]
This paper studies score approximation, estimation, and distribution recovery of diffusion models, when data are supported on an unknown low-dimensional linear subspace.
We show that with a properly chosen neural network architecture, the score function can be both accurately approximated and efficiently estimated.
The generated distribution based on the estimated score function captures the data geometric structures and converges to a close vicinity of the data distribution.
arXiv Detail & Related papers (2023-02-14T17:02:35Z) - Score-based Continuous-time Discrete Diffusion Models [102.65769839899315]
We extend diffusion models to discrete variables by introducing a Markov jump process where the reverse process denoises via a continuous-time Markov chain.
We show that an unbiased estimator can be obtained via simple matching the conditional marginal distributions.
We demonstrate the effectiveness of the proposed method on a set of synthetic and real-world music and image benchmarks.
arXiv Detail & Related papers (2022-11-30T05:33:29Z) - Variance-Aware Estimation of Kernel Mean Embedding [8.277998582564784]
We show how to speed-up convergence by leveraging variance information in the reproducing kernel Hilbert space.
We show that even when such information is a priori unknown, we can efficiently estimate it from the data.
arXiv Detail & Related papers (2022-10-13T01:58:06Z) - Continuous and Distribution-free Probabilistic Wind Power Forecasting: A
Conditional Normalizing Flow Approach [1.684864188596015]
We present a data-driven approach for probabilistic wind power forecasting based on conditional normalizing flow (CNF)
In contrast with the existing, this approach is distribution-free (as for non-parametric and quantile-based approaches) and can directly yield continuous probability densities.
arXiv Detail & Related papers (2022-06-06T08:48:58Z) - Distributional Gradient Boosting Machines [77.34726150561087]
Our framework is based on XGBoost and LightGBM.
We show that our framework achieves state-of-the-art forecast accuracy.
arXiv Detail & Related papers (2022-04-02T06:32:19Z) - Efficient CDF Approximations for Normalizing Flows [64.60846767084877]
We build upon the diffeomorphic properties of normalizing flows to estimate the cumulative distribution function (CDF) over a closed region.
Our experiments on popular flow architectures and UCI datasets show a marked improvement in sample efficiency as compared to traditional estimators.
arXiv Detail & Related papers (2022-02-23T06:11:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.