Efficient implementations of echo state network cross-validation
- URL: http://arxiv.org/abs/2006.11282v2
- Date: Thu, 3 Dec 2020 19:12:48 GMT
- Title: Efficient implementations of echo state network cross-validation
- Authors: Mantas Luko\v{s}evi\v{c}ius and Arnas Uselis
- Abstract summary: Cross-Validation (CV) is still uncommon in time series modeling.
We discuss CV of time series for predicting a concrete time interval of interest.
We introduce an efficient algorithm for implementing them.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Background/introduction: Cross-Validation (CV) is still uncommon in time
series modeling. Echo State Networks (ESNs), as a prime example of Reservoir
Computing (RC) models, are known for their fast and precise one-shot learning,
that often benefit from good hyper-parameter tuning. This makes them ideal to
change the status quo.
Methods: We discuss CV of time series for predicting a concrete time interval
of interest, suggest several schemes for cross-validating ESNs and introduce an
efficient algorithm for implementing them. This algorithm is presented as two
levels of optimizations of doing $k$-fold CV. Training an RC model typically
consists of two stages: (i) running the reservoir with the data and (ii)
computing the optimal readouts. The first level of our optimization addresses
the most computationally expensive part (i) and makes it remain constant
irrespective of $k$. It dramatically reduces reservoir computations in any type
of RC system and is enough if $k$ is small. The second level of optimization
also makes the (ii) part remain constant irrespective of large $k$, as long as
the dimension of the output is low. We discuss when the proposed validation
schemes for ESNs could be beneficial, three options for producing the final
model and empirically investigate them on six different real-world datasets, as
well as do empirical computation time experiments. We provide the code in an
online repository.
Results: Proposed CV schemes give better and more stable test performance in
all the six different real-world datasets, three task types. Empirical run
times confirm our complexity analysis.
Conclusions: In most situations $k$-fold CV of ESNs and many other RC models
can be done for virtually the same time and space complexity as a simple
single-split validation. This enables CV to become a standard practice in RC.
Related papers
- REP: Resource-Efficient Prompting for On-device Continual Learning [23.92661395403251]
On-device continual learning (CL) requires the co-optimization of model accuracy and resource efficiency to be practical.
It is commonly believed that CNN-based CL excels in resource efficiency, whereas ViT-based CL is superior in model performance.
We introduce REP, which improves resource efficiency specifically targeting prompt-based rehearsal-free methods.
arXiv Detail & Related papers (2024-06-07T09:17:33Z) - Provably Efficient CVaR RL in Low-rank MDPs [58.58570425202862]
We study risk-sensitive Reinforcement Learning (RL)
We propose a novel Upper Confidence Bound (UCB) bonus-driven algorithm to balance interplay between exploration, exploitation, and representation learning in CVaR RL.
We prove that our algorithm achieves a sample complexity of $epsilon$-optimal CVaR, where $H$ is the length of each episode, $A$ is the capacity of action space, and $d$ is the dimension of representations.
arXiv Detail & Related papers (2023-11-20T17:44:40Z) - A Specialized Semismooth Newton Method for Kernel-Based Optimal
Transport [92.96250725599958]
Kernel-based optimal transport (OT) estimators offer an alternative, functional estimation procedure to address OT problems from samples.
We show that our SSN method achieves a global convergence rate of $O (1/sqrtk)$, and a local quadratic convergence rate under standard regularity conditions.
arXiv Detail & Related papers (2023-10-21T18:48:45Z) - Improved Distribution Matching for Dataset Condensation [91.55972945798531]
We propose a novel dataset condensation method based on distribution matching.
Our simple yet effective method outperforms most previous optimization-oriented methods with much fewer computational resources.
arXiv Detail & Related papers (2023-07-19T04:07:33Z) - Finite-Time Analysis of Fully Decentralized Single-Timescale
Actor-Critic [4.94128206910124]
We introduce a fully decentralized Actor-Critic (AC) algorithm, where actor, critic, and global reward estimator are updated in an alternating manner.
We show that our algorithm has sample complexity of $tildemathcalO(epsilon-2)$ under Markovian sampling.
We also provide a local action privacy-preserving version of our algorithm and its analysis.
arXiv Detail & Related papers (2022-06-12T13:14:14Z) - An Experimental Design Perspective on Model-Based Reinforcement Learning [73.37942845983417]
In practical applications of RL, it is expensive to observe state transitions from the environment.
We propose an acquisition function that quantifies how much information a state-action pair would provide about the optimal solution to a Markov decision process.
arXiv Detail & Related papers (2021-12-09T23:13:57Z) - Targeted Cross-Validation [23.689101487016266]
We propose a targeted cross-validation (TCV) to select models or procedures based on a general weighted $L$ loss.
We show that the TCV is consistent in selecting the best performing candidate under the $L$ loss.
In this work, we broaden the concept of the selection consistency by allowing the best candidate to switch as the sample size varies.
arXiv Detail & Related papers (2021-09-14T19:53:18Z) - Leave Zero Out: Towards a No-Cross-Validation Approach for Model
Selection [21.06860861548758]
Cross Validation (CV) is the main workhorse for model selection.
CV suffers a conservatively biased estimation, since some part of the limited data has to hold out for validation.
CV tends to be extremely cumbersome, e.g., intolerant time-consuming, due to the repeated training procedures.
arXiv Detail & Related papers (2020-12-24T16:11:53Z) - AutoSimulate: (Quickly) Learning Synthetic Data Generation [70.82315853981838]
We propose an efficient alternative for optimal synthetic data generation based on a novel differentiable approximation of the objective.
We demonstrate that the proposed method finds the optimal data distribution faster (up to $50times$), with significantly reduced training data generation (up to $30times$) and better accuracy ($+8.7%$) on real-world test datasets than previous methods.
arXiv Detail & Related papers (2020-08-16T11:36:11Z) - Approximate Cross-Validation for Structured Models [20.79997929155929]
Gold standard evaluation technique is structured cross-validation (CV)
But CV here can be prohibitively slow due to the need to re-run already-expensive learning algorithms many times.
Previous work has shown approximate cross-validation (ACV) methods provide a fast and provably accurate alternative.
arXiv Detail & Related papers (2020-06-23T00:06:03Z) - The Right Tool for the Job: Matching Model and Instance Complexities [62.95183777679024]
As NLP models become larger, executing a trained model requires significant computational resources incurring monetary and environmental costs.
We propose a modification to contextual representation fine-tuning which, during inference, allows for an early (and fast) "exit"
We test our proposed modification on five different datasets in two tasks: three text classification datasets and two natural language inference benchmarks.
arXiv Detail & Related papers (2020-04-16T04:28:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.