A Novel Update Mechanism for Q-Networks Based On Extreme Learning
Machines
- URL: http://arxiv.org/abs/2006.02986v1
- Date: Thu, 4 Jun 2020 16:16:13 GMT
- Title: A Novel Update Mechanism for Q-Networks Based On Extreme Learning
Machines
- Authors: Callum Wilson, Annalisa Riccardi, Edmondo Minisci
- Abstract summary: Extreme Q-Learning Machine (EQLM) is applied to a reinforcement learning problem in the same manner as gradient based updates.
We compare its performance to a typical Q-Network on the cart-pole task.
We show EQLM has similar long-term learning performance to a Q-Network.
- Score: 0.6445605125467573
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Reinforcement learning is a popular machine learning paradigm which can find
near optimal solutions to complex problems. Most often, these procedures
involve function approximation using neural networks with gradient based
updates to optimise weights for the problem being considered. While this common
approach generally works well, there are other update mechanisms which are
largely unexplored in reinforcement learning. One such mechanism is Extreme
Learning Machines. These were initially proposed to drastically improve the
training speed of neural networks and have since seen many applications. Here
we attempt to apply extreme learning machines to a reinforcement learning
problem in the same manner as gradient based updates. This new algorithm is
called Extreme Q-Learning Machine (EQLM). We compare its performance to a
typical Q-Network on the cart-pole task - a benchmark reinforcement learning
problem - and show EQLM has similar long-term learning performance to a
Q-Network.
Related papers
- A Unified Framework for Neural Computation and Learning Over Time [56.44910327178975]
Hamiltonian Learning is a novel unified framework for learning with neural networks "over time"
It is based on differential equations that: (i) can be integrated without the need of external software solvers; (ii) generalize the well-established notion of gradient-based learning in feed-forward and recurrent networks; (iii) open to novel perspectives.
arXiv Detail & Related papers (2024-09-18T14:57:13Z) - Centered Self-Attention Layers [89.21791761168032]
The self-attention mechanism in transformers and the message-passing mechanism in graph neural networks are repeatedly applied.
We show that this application inevitably leads to oversmoothing, i.e., to similar representations at the deeper layers.
We present a correction term to the aggregating operator of these mechanisms.
arXiv Detail & Related papers (2023-06-02T15:19:08Z) - Deep learning applied to computational mechanics: A comprehensive
review, state of the art, and the classics [77.34726150561087]
Recent developments in artificial neural networks, particularly deep learning (DL), are reviewed in detail.
Both hybrid and pure machine learning (ML) methods are discussed.
History and limitations of AI are recounted and discussed, with particular attention at pointing out misstatements or misconceptions of the classics.
arXiv Detail & Related papers (2022-12-18T02:03:00Z) - Online Target Q-learning with Reverse Experience Replay: Efficiently
finding the Optimal Policy for Linear MDPs [50.75812033462294]
We bridge the gap between practical success of Q-learning and pessimistic theoretical results.
We present novel methods Q-Rex and Q-RexDaRe.
We show that Q-Rex efficiently finds the optimal policy for linear MDPs.
arXiv Detail & Related papers (2021-10-16T01:47:41Z) - Improving the sample-efficiency of neural architecture search with
reinforcement learning [0.0]
In this work, we would like to contribute to the area of Automated Machine Learning (AutoML)
Our focus is on one of the most promising research directions, reinforcement learning.
The validation accuracies of the child networks serve as a reward signal for training the controller.
We propose to modify this to a more modern and complex algorithm, PPO, which has demonstrated to be faster and more stable in other environments.
arXiv Detail & Related papers (2021-10-13T14:30:09Z) - Analytically Tractable Bayesian Deep Q-Learning [0.0]
We adapt the temporal difference Q-learning framework to make it compatible with the tractable approximate Gaussian inference (TAGI)
We demonstrate that TAGI can reach a performance comparable to backpropagation-trained networks.
arXiv Detail & Related papers (2021-06-21T13:11:52Z) - Neural Network-based Quantization for Network Automation [0.7034976835586089]
We introduce Bounding Sphere Quantization (BSQ) algorithm, a modification of the k-Means algorithm, that was shown to create better quantizations for certain network management use-cases.
BSQ required a significantly longer time to train than k-Means, a challenge which can be overcome with a neural network-based implementation.
We present such an implementation of BSQ that utilizes state-of-the-art deep learning tools to achieve a competitive training speed.
arXiv Detail & Related papers (2021-03-04T11:41:19Z) - Superiorities of Deep Extreme Learning Machines against Convolutional
Neural Networks [3.04585143845864]
Deep Learning (DL) is a machine learning procedure for artificial intelligence that analyzes the input data in detail.
DL has a popularity with the common improvements on the graphical processing unit capabilities.
Deep Extreme Learning machines (Deep ELM) is one of the fastest and effective way to meet fast classification problems.
arXiv Detail & Related papers (2021-01-21T08:22:18Z) - AutoML-Zero: Evolving Machine Learning Algorithms From Scratch [76.83052807776276]
We show that it is possible to automatically discover complete machine learning algorithms just using basic mathematical operations as building blocks.
We demonstrate this by introducing a novel framework that significantly reduces human bias through a generic search space.
We believe these preliminary successes in discovering machine learning algorithms from scratch indicate a promising new direction in the field.
arXiv Detail & Related papers (2020-03-06T19:00:04Z) - Deep Learning for Ultra-Reliable and Low-Latency Communications in 6G
Networks [84.2155885234293]
We first summarize how to apply data-driven supervised deep learning and deep reinforcement learning in URLLC.
To address these open problems, we develop a multi-level architecture that enables device intelligence, edge intelligence, and cloud intelligence for URLLC.
arXiv Detail & Related papers (2020-02-22T14:38:11Z) - Memristor Hardware-Friendly Reinforcement Learning [14.853739554366351]
We propose a memristive neuromorphic hardware implementation for the actor-critic algorithm in reinforcement learning.
We consider the task of balancing an inverted pendulum, a classical problem in both RL and control theory.
We believe that this study shows the promise of using memristor-based hardware neural networks for handling complex tasks through in-situ reinforcement learning.
arXiv Detail & Related papers (2020-01-20T01:08:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.