Statistical learning and cross-validation for point processes
- URL: http://arxiv.org/abs/2103.01356v1
- Date: Mon, 1 Mar 2021 23:47:48 GMT
- Title: Statistical learning and cross-validation for point processes
- Authors: Ottmar Cronie, Mehdi Moradi, Christophe A.N. Biscio
- Abstract summary: This paper presents the first general (parametric) statistical learning framework for point processes in general spaces.
The general idea is to carry out the fitting by predicting CV-generated validation sets using the corresponding training sets.
We numerically show that our statistical learning approach outperforms the state of the art in terms of mean (integrated) squared error.
- Score: 0.9281671380673306
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper presents the first general (supervised) statistical learning
framework for point processes in general spaces. Our approach is based on the
combination of two new concepts, which we define in the paper: i) bivariate
innovations, which are measures of discrepancy/prediction-accuracy between two
point processes, and ii) point process cross-validation (CV), which we here
define through point process thinning. The general idea is to carry out the
fitting by predicting CV-generated validation sets using the corresponding
training sets; the prediction error, which we minimise, is measured by means of
bivariate innovations. Having established various theoretical properties of our
bivariate innovations, we study in detail the case where the CV procedure is
obtained through independent thinning and we apply our statistical learning
methodology to three typical spatial statistical settings, namely parametric
intensity estimation, non-parametric intensity estimation and Papangelou
conditional intensity fitting. Aside from deriving theoretical properties
related to these cases, in each of them we numerically show that our
statistical learning approach outperforms the state of the art in terms of mean
(integrated) squared error.
Related papers
- Targeted Learning for Variable Importance [23.428985354228672]
variable importance is one of the most widely used measures for interpreting machine learning.
We propose a novel method by employing the targeted learning (TL) framework, designed to enhance robustness in inference for variable importance metrics.
We show that it (i) retains the efficiency of traditional methods, (ii) maintains comparable computational complexity, and (iii) delivers improved accuracy, especially in finite sample contexts.
arXiv Detail & Related papers (2024-11-04T16:14:45Z) - Statistical Inference for Temporal Difference Learning with Linear Function Approximation [62.69448336714418]
Temporal Difference (TD) learning, arguably the most widely used for policy evaluation, serves as a natural framework for this purpose.
In this paper, we study the consistency properties of TD learning with Polyak-Ruppert averaging and linear function approximation, and obtain three significant improvements over existing results.
arXiv Detail & Related papers (2024-10-21T15:34:44Z) - Exogenous Matching: Learning Good Proposals for Tractable Counterfactual Estimation [1.9662978733004601]
We propose an importance sampling method for tractable and efficient estimation of counterfactual expressions.
By minimizing a common upper bound of counterfactual estimators, we transform the variance minimization problem into a conditional distribution learning problem.
We validate the theoretical results through experiments under various types and settings of Structural Causal Models (SCMs) and demonstrate the outperformance on counterfactual estimation tasks.
arXiv Detail & Related papers (2024-10-17T03:08:28Z) - Selective Nonparametric Regression via Testing [54.20569354303575]
We develop an abstention procedure via testing the hypothesis on the value of the conditional variance at a given point.
Unlike existing methods, the proposed one allows to account not only for the value of the variance itself but also for the uncertainty of the corresponding variance predictor.
arXiv Detail & Related papers (2023-09-28T13:04:11Z) - Variational Distillation for Multi-View Learning [104.17551354374821]
We design several variational information bottlenecks to exploit two key characteristics for multi-view representation learning.
Under rigorously theoretical guarantee, our approach enables IB to grasp the intrinsic correlation between observations and semantic labels.
arXiv Detail & Related papers (2022-06-20T03:09:46Z) - Regularizing Variational Autoencoder with Diversity and Uncertainty
Awareness [61.827054365139645]
Variational Autoencoder (VAE) approximates the posterior of latent variables based on amortized variational inference.
We propose an alternative model, DU-VAE, for learning a more Diverse and less Uncertain latent space.
arXiv Detail & Related papers (2021-10-24T07:58:13Z) - Learning Prediction Intervals for Regression: Generalization and
Calibration [12.576284277353606]
We study the generation of prediction intervals in regression for uncertainty quantification.
We use a general learning theory to characterize the optimality-feasibility tradeoff that encompasses Lipschitz continuity and VC-subgraph classes.
We empirically demonstrate the strengths of our interval generation and calibration algorithms in terms of testing performances compared to existing benchmarks.
arXiv Detail & Related papers (2021-02-26T17:55:30Z) - Counterfactual Representation Learning with Balancing Weights [74.67296491574318]
Key to causal inference with observational data is achieving balance in predictive features associated with each treatment type.
Recent literature has explored representation learning to achieve this goal.
We develop an algorithm for flexible, scalable and accurate estimation of causal effects.
arXiv Detail & Related papers (2020-10-23T19:06:03Z) - Statistical Inference for Networks of High-Dimensional Point Processes [19.38934705817528]
We develop a new statistical inference procedure for high-dimensional Hawkes processes.
The key ingredient for this inference procedure is a new concentration inequality on the first- and second-order statistics.
We demonstrate their utility by applying them to a neuron spike train data set.
arXiv Detail & Related papers (2020-07-15T02:46:36Z) - A One-step Approach to Covariate Shift Adaptation [82.01909503235385]
A default assumption in many machine learning scenarios is that the training and test samples are drawn from the same probability distribution.
We propose a novel one-step approach that jointly learns the predictive model and the associated weights in one optimization.
arXiv Detail & Related papers (2020-07-08T11:35:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.