Stochastic Approximation for High-frequency Observations in Data
Assimilation
- URL: http://arxiv.org/abs/2011.02672v1
- Date: Thu, 5 Nov 2020 06:02:27 GMT
- Title: Stochastic Approximation for High-frequency Observations in Data
Assimilation
- Authors: Shushu Zhang, Vivak Patel
- Abstract summary: High-frequency sensors offer opportunities for higher statistical accuracy of down-stream estimates, but their frequency results in a plethora of computational problems in data assimilation tasks.
We adapt approximation methods to address the unique challenges of high-frequency observations in data assimilation.
As a result, we are able to produce estimates that leverage all of the observations in a manner that avoids the aforementioned computational problems and preserves the statistical accuracy of the estimates.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the increasing penetration of high-frequency sensors across a number of
biological and physical systems, the abundance of the resulting observations
offers opportunities for higher statistical accuracy of down-stream estimates,
but their frequency results in a plethora of computational problems in data
assimilation tasks. The high-frequency of these observations has been
traditionally dealt with by using data modification strategies such as
accumulation, averaging, and sampling. However, these data modification
strategies will reduce the quality of the estimates, which may be untenable for
many systems. Therefore, to ensure high-quality estimates, we adapt stochastic
approximation methods to address the unique challenges of high-frequency
observations in data assimilation. As a result, we are able to produce
estimates that leverage all of the observations in a manner that avoids the
aforementioned computational problems and preserves the statistical accuracy of
the estimates.
Related papers
- Denoising ESG: quantifying data uncertainty from missing data with Machine Learning and prediction intervals [43.5280757929682]
This paper explores the application of established machine learning techniques for imputing missing data in a real-world ESG dataset.
By employing multiple imputation strategies, this study assesses the robustness of imputation methods and quantifies the uncertainty associated with missing data.
arXiv Detail & Related papers (2024-07-29T14:31:44Z) - Harnessing Data Augmentation to Quantify Uncertainty in the Early
Estimation of Single-Photon Source Quality [8.397730500554047]
This study investigates the use of data augmentation, a machine learning technique, to supplement experimental data with bootstrapped samples.
Eight datasets obtained from measurements involving a single InGaAs/GaAs epitaxial quantum dot serve as a proof-of-principle example.
arXiv Detail & Related papers (2023-06-22T05:17:26Z) - Anomaly Detection with Variance Stabilized Density Estimation [49.46356430493534]
We present a variance-stabilized density estimation problem for maximizing the likelihood of the observed samples.
To obtain a reliable anomaly detector, we introduce a spectral ensemble of autoregressive models for learning the variance-stabilized distribution.
We have conducted an extensive benchmark with 52 datasets, demonstrating that our method leads to state-of-the-art results.
arXiv Detail & Related papers (2023-06-01T11:52:58Z) - DeepVol: Volatility Forecasting from High-Frequency Data with Dilated Causal Convolutions [53.37679435230207]
We propose DeepVol, a model based on Dilated Causal Convolutions that uses high-frequency data to forecast day-ahead volatility.
Our empirical results suggest that the proposed deep learning-based approach effectively learns global features from high-frequency data.
arXiv Detail & Related papers (2022-09-23T16:13:47Z) - Scalable Sensitivity and Uncertainty Analysis for Causal-Effect
Estimates of Continuous-Valued Interventions [34.19821413853115]
Estimating the effects of continuous-valued interventions from observational data is critically important in fields such as climate science, healthcare, and economics.
We develop a continuous treatment-effect marginal sensitivity model (CMSM) and derive bounds that agree with both the observed data and a researcher-defined level of hidden confounding.
We introduce a scalable algorithm to derive the bounds and uncertainty-aware deep models to efficiently estimate these bounds for high-dimensional, large-sample observational data.
arXiv Detail & Related papers (2022-04-21T11:15:10Z) - High Dimensional Statistical Estimation under One-bit Quantization [27.718986773043643]
One-bit (binary) data are preferable in many applications because of the efficiency in signal storage, processing, transmission, and enhancement of privacy.
In this paper, we study three fundamental statistical estimation problems.
Under both sub-Gaussian and heavy-tailed regimes, new estimators that handle high-dimensional scaling are proposed.
arXiv Detail & Related papers (2022-02-26T15:13:04Z) - Gaining Outlier Resistance with Progressive Quantiles: Fast Algorithms
and Theoretical Studies [1.6457778420360534]
A framework of outlier-resistant estimation is introduced to robustify arbitrarily loss function.
A new technique is proposed to alleviate the requirement on starting point such that on regular datasets the number of data reestimations can be substantially reduced.
The obtained estimators, though not necessarily globally or even globally, enjoymax optimality in both low dimensions.
arXiv Detail & Related papers (2021-12-15T20:35:21Z) - Doing Great at Estimating CATE? On the Neglected Assumptions in
Benchmark Comparisons of Treatment Effect Estimators [91.3755431537592]
We show that even in arguably the simplest setting, estimation under ignorability assumptions can be misleading.
We consider two popular machine learning benchmark datasets for evaluation of heterogeneous treatment effect estimators.
We highlight that the inherent characteristics of the benchmark datasets favor some algorithms over others.
arXiv Detail & Related papers (2021-07-28T13:21:27Z) - Quantifying Uncertainty in Deep Spatiotemporal Forecasting [67.77102283276409]
We describe two types of forecasting problems: regular grid-based and graph-based.
We analyze UQ methods from both the Bayesian and the frequentist point view, casting in a unified framework via statistical decision theory.
Through extensive experiments on real-world road network traffic, epidemics, and air quality forecasting tasks, we reveal the statistical computational trade-offs for different UQ methods.
arXiv Detail & Related papers (2021-05-25T14:35:46Z) - Variable Skipping for Autoregressive Range Density Estimation [84.60428050170687]
We show a technique, variable skipping, for accelerating range density estimation over deep autoregressive models.
We show that variable skipping provides 10-100$times$ efficiency improvements when targeting challenging high-quantile error metrics.
arXiv Detail & Related papers (2020-07-10T19:01:40Z) - Unlabelled Data Improves Bayesian Uncertainty Calibration under
Covariate Shift [100.52588638477862]
We develop an approximate Bayesian inference scheme based on posterior regularisation.
We demonstrate the utility of our method in the context of transferring prognostic models of prostate cancer across globally diverse populations.
arXiv Detail & Related papers (2020-06-26T13:50:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.