Hyperparameter Tuning in Echo State Networks
- URL: http://arxiv.org/abs/2207.07976v1
- Date: Sat, 16 Jul 2022 16:20:01 GMT
- Title: Hyperparameter Tuning in Echo State Networks
- Authors: Filip Matzner
- Abstract summary: Echo State Networks are a type of recurrent neural network with a large randomly generated reservoir and a small number of readout connections trained via linear regression.
We propose an alternative approach of hyper parameter tuning based on the Covariance Matrix Adaptation Evolution Strategy (CMA-ES)
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Echo State Networks represent a type of recurrent neural network with a large
randomly generated reservoir and a small number of readout connections trained
via linear regression. The most common topology of the reservoir is a fully
connected network of up to thousands of neurons. Over the years, researchers
have introduced a variety of alternative reservoir topologies, such as a
circular network or a linear path of connections. When comparing the
performance of different topologies or other architectural changes, it is
necessary to tune the hyperparameters for each of the topologies separately
since their properties may significantly differ. The hyperparameter tuning is
usually carried out manually by selecting the best performing set of parameters
from a sparse grid of predefined combinations. Unfortunately, this approach may
lead to underperforming configurations, especially for sensitive topologies. We
propose an alternative approach of hyperparameter tuning based on the
Covariance Matrix Adaptation Evolution Strategy (CMA-ES). Using this approach,
we have improved multiple topology comparison results by orders of magnitude
suggesting that topology alone does not play as important role as properly
tuned hyperparameters.
Related papers
- Optimal Recurrent Network Topologies for Dynamical Systems Reconstruction [5.498605900447094]
We seek to infer from time series measurements a generative model of the underlying dynamical process.
A common strategy here is parameter pruning, removing all parameters with small weights.
We show that geometric pruning, where in contrast to magnitude-based pruning weights with a low contribution to an attractor's geometrical structure are removed, indeed manages to reduce parameter load substantially.
arXiv Detail & Related papers (2024-06-07T13:41:17Z) - Structurally Flexible Neural Networks: Evolving the Building Blocks for General Agents [7.479827648985631]
We show that it is possible to optimize a set of distinct neuron- and synapse types allowing for a mitigation of the symmetry dilemma.
We demonstrate this by optimizing a single set of neurons and synapses to solve multiple reinforcement learning control tasks simultaneously.
arXiv Detail & Related papers (2024-04-06T14:04:14Z) - Principled Architecture-aware Scaling of Hyperparameters [69.98414153320894]
Training a high-quality deep neural network requires choosing suitable hyperparameters, which is a non-trivial and expensive process.
In this work, we precisely characterize the dependence of initializations and maximal learning rates on the network architecture.
We demonstrate that network rankings can be easily changed by better training networks in benchmarks.
arXiv Detail & Related papers (2024-02-27T11:52:49Z) - Learning to Learn with Generative Models of Neural Network Checkpoints [71.06722933442956]
We construct a dataset of neural network checkpoints and train a generative model on the parameters.
We find that our approach successfully generates parameters for a wide range of loss prompts.
We apply our method to different neural network architectures and tasks in supervised and reinforcement learning.
arXiv Detail & Related papers (2022-09-26T17:59:58Z) - AUTOMATA: Gradient Based Data Subset Selection for Compute-Efficient
Hyper-parameter Tuning [72.54359545547904]
We propose a gradient-based subset selection framework for hyper- parameter tuning.
We show that using gradient-based data subsets for hyper- parameter tuning achieves significantly faster turnaround times and speedups of 3$times$-30$times$.
arXiv Detail & Related papers (2022-03-15T19:25:01Z) - Lightweight Convolutional Neural Networks By Hypercomplex
Parameterization [10.420215908252425]
We define the parameterization of hypercomplex convolutional layers to develop lightweight and efficient large-scale convolutional models.
Our method grasps the convolution rules and the filters organization directly from data.
We demonstrate the versatility of this approach to multiple domains of application by performing experiments on various image datasets and audio datasets.
arXiv Detail & Related papers (2021-10-08T14:57:19Z) - Manifold Regularized Dynamic Network Pruning [102.24146031250034]
This paper proposes a new paradigm that dynamically removes redundant filters by embedding the manifold information of all instances into the space of pruned networks.
The effectiveness of the proposed method is verified on several benchmarks, which shows better performance in terms of both accuracy and computational cost.
arXiv Detail & Related papers (2021-03-10T03:59:03Z) - Online hyperparameter optimization by real-time recurrent learning [57.01871583756586]
Our framework takes advantage of the analogy between hyperparameter optimization and parameter learning in neural networks (RNNs)
It adapts a well-studied family of online learning algorithms for RNNs to tune hyperparameters and network parameters simultaneously.
This procedure yields systematically better generalization performance compared to standard methods, at a fraction of wallclock time.
arXiv Detail & Related papers (2021-02-15T19:36:18Z) - A Study of Genetic Algorithms for Hyperparameter Optimization of Neural
Networks in Machine Translation [0.0]
We propose an automatic tuning method modeled after Darwin's Survival of the Fittest Theory via a Genetic Algorithm.
Research results show that the proposed method, a GA, outperforms a random selection of hyper parameters.
arXiv Detail & Related papers (2020-09-15T02:24:16Z) - Automatic Setting of DNN Hyper-Parameters by Mixing Bayesian
Optimization and Tuning Rules [0.6875312133832078]
We build a new algorithm for evaluating and analyzing the results of the network on the training and validation sets.
We use a set of tuning rules to add new hyper-parameters and/or to reduce the hyper- parameter search space to select a better combination.
arXiv Detail & Related papers (2020-06-03T08:53:48Z) - Model Fusion via Optimal Transport [64.13185244219353]
We present a layer-wise model fusion algorithm for neural networks.
We show that this can successfully yield "one-shot" knowledge transfer between neural networks trained on heterogeneous non-i.i.d. data.
arXiv Detail & Related papers (2019-10-12T22:07:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.