Scalable Real-Time Recurrent Learning Using Columnar-Constructive
Networks
- URL: http://arxiv.org/abs/2302.05326v3
- Date: Tue, 21 Nov 2023 19:33:57 GMT
- Title: Scalable Real-Time Recurrent Learning Using Columnar-Constructive
Networks
- Authors: Khurram Javed, Haseeb Shah, Rich Sutton, Martha White
- Abstract summary: We propose two constraints that make real-time recurrent learning scalable.
We show that by either decomposing the network into independent modules or learning the network in stages, we can make RTRL scale linearly with the number of parameters.
We demonstrate the effectiveness of our approach over Truncated-BPTT on a prediction benchmark inspired by animal learning and by doing policy evaluation of pre-trained policies for Atari 2600 games.
- Score: 19.248060562241296
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Constructing states from sequences of observations is an important component
of reinforcement learning agents. One solution for state construction is to use
recurrent neural networks. Back-propagation through time (BPTT), and real-time
recurrent learning (RTRL) are two popular gradient-based methods for recurrent
learning. BPTT requires complete trajectories of observations before it can
compute the gradients and is unsuitable for online updates. RTRL can do online
updates but scales poorly to large networks. In this paper, we propose two
constraints that make RTRL scalable. We show that by either decomposing the
network into independent modules or learning the network in stages, we can make
RTRL scale linearly with the number of parameters. Unlike prior scalable
gradient estimation algorithms, such as UORO and Truncated-BPTT, our algorithms
do not add noise or bias to the gradient estimate. Instead, they trade off the
functional capacity of the network for computationally efficient learning. We
demonstrate the effectiveness of our approach over Truncated-BPTT on a
prediction benchmark inspired by animal learning and by doing policy evaluation
of pre-trained policies for Atari 2600 games.
Related papers
- Gradient-Free Training of Recurrent Neural Networks using Random Perturbations [1.1742364055094265]
Recurrent neural networks (RNNs) hold immense potential for computations due to their Turing completeness and sequential processing capabilities.
Backpropagation through time (BPTT), the prevailing method, extends the backpropagation algorithm by unrolling the RNN over time.
BPTT suffers from significant drawbacks, including the need to interleave forward and backward phases and store exact gradient information.
We present a new approach to perturbation-based learning in RNNs whose performance is competitive with BPTT.
arXiv Detail & Related papers (2024-05-14T21:15:29Z) - Real-Time Recurrent Reinforcement Learning [7.737685867200335]
RTRRL consists of three parts: (1) a Meta-RL RNN architecture, implementing on its own an actor-critic algorithm; (2) an outer reinforcement learning algorithm, exploiting temporal difference learning and dutch eligibility traces to train the Meta-RL network; and (3) random-feedback local-online (RFLO) learning, an online automatic differentiation algorithm for computing the gradients with respect to parameters of the network.
arXiv Detail & Related papers (2023-11-08T16:56:16Z) - Efficient Real Time Recurrent Learning through combined activity and
parameter sparsity [0.5076419064097732]
Backpropagation through time (BPTT) is the standard algorithm for training recurrent neural networks (RNNs)
BPTT is unsuited for online learning and presents a challenge for implementation on low-resource real-time systems.
We show that recurrent networks exhibiting high activity sparsity can reduce the computational cost of Real-Time Recurrent Learning (RTRL)
arXiv Detail & Related papers (2023-03-10T01:09:04Z) - Towards Memory- and Time-Efficient Backpropagation for Training Spiking
Neural Networks [70.75043144299168]
Spiking Neural Networks (SNNs) are promising energy-efficient models for neuromorphic computing.
We propose the Spatial Learning Through Time (SLTT) method that can achieve high performance while greatly improving training efficiency.
Our method achieves state-of-the-art accuracy on ImageNet, while the memory cost and training time are reduced by more than 70% and 50%, respectively, compared with BPTT.
arXiv Detail & Related papers (2023-02-28T05:01:01Z) - Boosting Offline Reinforcement Learning via Data Rebalancing [104.3767045977716]
offline reinforcement learning (RL) is challenged by the distributional shift between learning policies and datasets.
We propose a simple yet effective method to boost offline RL algorithms based on the observation that resampling a dataset keeps the distribution support unchanged.
We dub our method ReD (Return-based Data Rebalance), which can be implemented with less than 10 lines of code change and adds negligible running time.
arXiv Detail & Related papers (2022-10-17T16:34:01Z) - Online Training Through Time for Spiking Neural Networks [66.7744060103562]
Spiking neural networks (SNNs) are promising brain-inspired energy-efficient models.
Recent progress in training methods has enabled successful deep SNNs on large-scale tasks with low latency.
We propose online training through time (OTTT) for SNNs, which is derived from BPTT to enable forward-in-time learning.
arXiv Detail & Related papers (2022-10-09T07:47:56Z) - Scalable Online Recurrent Learning Using Columnar Neural Networks [35.584855852204385]
An algorithm called RTRL can compute gradients for recurrent networks online but is computationally intractable for large networks.
We propose a credit-assignment algorithm that approximates the gradients for recurrent learning in real-time using $O(n)$ operations and memory per-step.
arXiv Detail & Related papers (2021-03-09T23:45:13Z) - Improving Computational Efficiency in Visual Reinforcement Learning via
Stored Embeddings [89.63764845984076]
We present Stored Embeddings for Efficient Reinforcement Learning (SEER)
SEER is a simple modification of existing off-policy deep reinforcement learning methods.
We show that SEER does not degrade the performance of RLizable agents while significantly saving computation and memory.
arXiv Detail & Related papers (2021-03-04T08:14:10Z) - RIFLE: Backpropagation in Depth for Deep Transfer Learning through
Re-Initializing the Fully-connected LayEr [60.07531696857743]
Fine-tuning the deep convolution neural network(CNN) using a pre-trained model helps transfer knowledge learned from larger datasets to the target task.
We propose RIFLE - a strategy that deepens backpropagation in transfer learning settings.
RIFLE brings meaningful updates to the weights of deep CNN layers and improves low-level feature learning.
arXiv Detail & Related papers (2020-07-07T11:27:43Z) - A Practical Sparse Approximation for Real Time Recurrent Learning [38.19296522866088]
Real Time Recurrent Learning (RTRL) eliminates the need for history storage and allows for online weight updates.
We introduce the Sparse n-step Approximation (SnAp) to the RTRL influence matrix, which only keeps entries that are nonzero within n steps of the recurrent core.
For highly sparse networks, SnAp with n=2 remains tractable and can outperform backpropagation through time in terms of learning speed when updates are done online.
arXiv Detail & Related papers (2020-06-12T14:38:15Z) - Large-Scale Gradient-Free Deep Learning with Recursive Local
Representation Alignment [84.57874289554839]
Training deep neural networks on large-scale datasets requires significant hardware resources.
Backpropagation, the workhorse for training these networks, is an inherently sequential process that is difficult to parallelize.
We propose a neuro-biologically-plausible alternative to backprop that can be used to train deep networks.
arXiv Detail & Related papers (2020-02-10T16:20:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.