Replay-enhanced Continual Reinforcement Learning
- URL: http://arxiv.org/abs/2311.11557v1
- Date: Mon, 20 Nov 2023 06:21:52 GMT
- Title: Replay-enhanced Continual Reinforcement Learning
- Authors: Tiantian Zhang, Kevin Zehua Shen, Zichuan Lin, Bo Yuan, Xueqian Wang,
Xiu Li, Deheng Ye
- Abstract summary: We introduce RECALL, a replay-enhanced method that greatly improves the plasticity of existing replay-based methods on new tasks.
Experiments on the Continual World benchmark show that RECALL performs significantly better than purely perfect memory replay.
- Score: 37.34722105058351
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Replaying past experiences has proven to be a highly effective approach for
averting catastrophic forgetting in supervised continual learning. However,
some crucial factors are still largely ignored, making it vulnerable to serious
failure, when used as a solution to forgetting in continual reinforcement
learning, even in the context of perfect memory where all data of previous
tasks are accessible in the current task. On the one hand, since most
reinforcement learning algorithms are not invariant to the reward scale, the
previously well-learned tasks (with high rewards) may appear to be more salient
to the current learning process than the current task (with small initial
rewards). This causes the agent to concentrate on those salient tasks at the
expense of generality on the current task. On the other hand, offline learning
on replayed tasks while learning a new task may induce a distributional shift
between the dataset and the learned policy on old tasks, resulting in
forgetting. In this paper, we introduce RECALL, a replay-enhanced method that
greatly improves the plasticity of existing replay-based methods on new tasks
while effectively avoiding the recurrence of catastrophic forgetting in
continual reinforcement learning. RECALL leverages adaptive normalization on
approximate targets and policy distillation on old tasks to enhance generality
and stability, respectively. Extensive experiments on the Continual World
benchmark show that RECALL performs significantly better than purely perfect
memory replay, and achieves comparable or better overall performance against
state-of-the-art continual learning methods.
Related papers
- Class-Incremental Learning Using Generative Experience Replay Based on
Time-aware Regularization [24.143811670210546]
Generative experience replay addresses the challenge of learning new tasks accumulatively without forgetting.
We introduce a time-aware regularization method to fine-tune the three training objective terms used for generative replay.
Experimental results indicate that our method pushes the limit of brain-inspired continual learners under such strict settings.
arXiv Detail & Related papers (2023-10-05T21:07:45Z) - Towards Robust Continual Learning with Bayesian Adaptive Moment Regularization [51.34904967046097]
Continual learning seeks to overcome the challenge of catastrophic forgetting, where a model forgets previously learnt information.
We introduce a novel prior-based method that better constrains parameter growth, reducing catastrophic forgetting.
Results show that BAdam achieves state-of-the-art performance for prior-based methods on challenging single-headed class-incremental experiments.
arXiv Detail & Related papers (2023-09-15T17:10:51Z) - AdaER: An Adaptive Experience Replay Approach for Continual Lifelong
Learning [16.457330925212606]
We present adaptive-experience replay (AdaER) to address the challenge of continual lifelong learning.
AdaER consists of two stages: memory replay and memory update.
Results: AdaER outperforms existing continual lifelong learning baselines.
arXiv Detail & Related papers (2023-08-07T01:25:45Z) - Detachedly Learn a Classifier for Class-Incremental Learning [11.865788374587734]
We present an analysis that the failure of vanilla experience replay (ER) comes from unnecessary re-learning of previous tasks and incompetence to distinguish current task from the previous ones.
We propose a novel replay strategy task-aware experience replay.
Experimental results show our method outperforms current state-of-the-art methods.
arXiv Detail & Related papers (2023-02-23T01:35:44Z) - Relational Experience Replay: Continual Learning by Adaptively Tuning
Task-wise Relationship [54.73817402934303]
We propose Experience Continual Replay (ERR), a bi-level learning framework to adaptively tune task-wise to achieve a better stability plasticity' tradeoff.
ERR can consistently improve the performance of all baselines and surpass current state-of-the-art methods.
arXiv Detail & Related papers (2021-12-31T12:05:22Z) - Center Loss Regularization for Continual Learning [0.0]
In general, neural networks lack the ability to learn different tasks sequentially.
Our approach remembers old tasks by projecting the representations of new tasks close to that of old tasks.
We demonstrate that our approach is scalable, effective, and gives competitive performance compared to state-of-the-art continual learning methods.
arXiv Detail & Related papers (2021-10-21T17:46:44Z) - Bilevel Continual Learning [76.50127663309604]
We present a novel framework of continual learning named "Bilevel Continual Learning" (BCL)
Our experiments on continual learning benchmarks demonstrate the efficacy of the proposed BCL compared to many state-of-the-art methods.
arXiv Detail & Related papers (2020-07-30T16:00:23Z) - Continual Deep Learning by Functional Regularisation of Memorable Past [95.97578574330934]
Continually learning new skills is important for intelligent systems, yet standard deep learning methods suffer from catastrophic forgetting of the past.
We propose a new functional-regularisation approach that utilises a few memorable past examples crucial to avoid forgetting.
Our method achieves state-of-the-art performance on standard benchmarks and opens a new direction for life-long learning where regularisation and memory-based methods are naturally combined.
arXiv Detail & Related papers (2020-04-29T10:47:54Z) - Generalized Hindsight for Reinforcement Learning [154.0545226284078]
We argue that low-reward data collected while trying to solve one task provides little to no signal for solving that particular task.
We present Generalized Hindsight: an approximate inverse reinforcement learning technique for relabeling behaviors with the right tasks.
arXiv Detail & Related papers (2020-02-26T18:57:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.