Deep learning for $\psi$-weakly dependent processes
- URL: http://arxiv.org/abs/2302.00333v1
- Date: Wed, 1 Feb 2023 09:31:15 GMT
- Title: Deep learning for $\psi$-weakly dependent processes
- Authors: William Kengne, Wade Modou
- Abstract summary: We perform deep neural networks for learning $psi$-weakly dependent processes.
The consistency of the empirical risk minimization algorithm in the class of deep neural networks predictors is established.
Some simulation results are provided, as well as an application to the US recession data.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we perform deep neural networks for learning $\psi$-weakly
dependent processes. Such weak-dependence property includes a class of weak
dependence conditions such as mixing, association,$\cdots$ and the setting
considered here covers many commonly used situations such as: regression
estimation, time series prediction, time series classification,$\cdots$ The
consistency of the empirical risk minimization algorithm in the class of deep
neural networks predictors is established. We achieve the generalization bound
and obtain a learning rate, which is less than $\mathcal{O}(n^{-1/\alpha})$,
for all $\alpha > 2 $. Applications to binary time series classification and
prediction in affine causal models with exogenous covariates are carried out.
Some simulation results are provided, as well as an application to the US
recession data.
Related papers
- Agnostic Smoothed Online Learning [5.167069404528051]
We propose an algorithm to guarantee sublinear regret for smoothed online learning without prior knowledge of $mu$.
R-Cover has adaptive regret $tilde O(sqrtdT/sigma)$ for function classes with dimension $d$, which is optimal up to logarithmic factors.
arXiv Detail & Related papers (2024-10-07T15:25:21Z) - Penalized deep neural networks estimator with general loss functions
under weak dependence [0.0]
This paper carries out sparse-penalized deep neural networks predictors for learning weakly dependent processes.
Some simulation results are provided, and application to the forecast of the particulate matter in the Vit'oria metropolitan area is also considered.
arXiv Detail & Related papers (2023-05-10T15:06:53Z) - Continuous time recurrent neural networks: overview and application to
forecasting blood glucose in the intensive care unit [56.801856519460465]
Continuous time autoregressive recurrent neural networks (CTRNNs) are a deep learning model that account for irregular observations.
We demonstrate the application of these models to probabilistic forecasting of blood glucose in a critical care setting.
arXiv Detail & Related papers (2023-04-14T09:39:06Z) - Sparse-penalized deep neural networks estimator under weak dependence [0.0]
We consider the nonparametric regression and the classification problems for $psi$-weakly dependent processes.
A penalized estimation method for sparse deep neural networks is performed.
arXiv Detail & Related papers (2023-03-02T16:53:51Z) - Excess risk bound for deep learning under weak dependence [0.0]
This paper considers deep neural networks for learning weakly dependent processes.
We derive the required depth, width and sparsity of a deep neural network to approximate any H"older smooth function.
arXiv Detail & Related papers (2023-02-15T07:23:48Z) - Neural Dependencies Emerging from Learning Massive Categories [94.77992221690742]
This work presents two astonishing findings on neural networks learned for large-scale image classification.
1) Given a well-trained model, the logits predicted for some category can be directly obtained by linearly combining the predictions of a few other categories.
2) Neural dependencies exist not only within a single model, but even between two independently learned models.
arXiv Detail & Related papers (2022-11-21T09:42:15Z) - The Separation Capacity of Random Neural Networks [78.25060223808936]
We show that a sufficiently large two-layer ReLU-network with standard Gaussian weights and uniformly distributed biases can solve this problem with high probability.
We quantify the relevant structure of the data in terms of a novel notion of mutual complexity.
arXiv Detail & Related papers (2021-07-31T10:25:26Z) - Locality defeats the curse of dimensionality in convolutional
teacher-student scenarios [69.2027612631023]
We show that locality is key in determining the learning curve exponent $beta$.
We conclude by proving, using a natural assumption, that performing kernel regression with a ridge that decreases with the size of the training set leads to similar learning curve exponents to those we obtain in the ridgeless case.
arXiv Detail & Related papers (2021-06-16T08:27:31Z) - Understanding the Under-Coverage Bias in Uncertainty Estimation [58.03725169462616]
quantile regression tends to emphunder-cover than the desired coverage level in reality.
We prove that quantile regression suffers from an inherent under-coverage bias.
Our theory reveals that this under-coverage bias stems from a certain high-dimensional parameter estimation error.
arXiv Detail & Related papers (2021-06-10T06:11:55Z) - Towards an Understanding of Benign Overfitting in Neural Networks [104.2956323934544]
Modern machine learning models often employ a huge number of parameters and are typically optimized to have zero training loss.
We examine how these benign overfitting phenomena occur in a two-layer neural network setting.
We show that it is possible for the two-layer ReLU network interpolator to achieve a near minimax-optimal learning rate.
arXiv Detail & Related papers (2021-06-06T19:08:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.