On Catastrophic Interference in Atari 2600 Games
- URL: http://arxiv.org/abs/2002.12499v2
- Date: Tue, 9 Jun 2020 17:36:46 GMT
- Title: On Catastrophic Interference in Atari 2600 Games
- Authors: William Fedus, Dibya Ghosh, John D. Martin, Marc G. Bellemare, Yoshua
Bengio, Hugo Larochelle
- Abstract summary: We show that interference causes performance to plateau.
We demonstrate performance boosts across architectures, learning algorithms and environments.
A more refined analysis shows that learning one segment of a game often increases prediction errors elsewhere.
- Score: 104.61596014400892
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Model-free deep reinforcement learning is sample inefficient. One hypothesis
-- speculated, but not confirmed -- is that catastrophic interference within an
environment inhibits learning. We test this hypothesis through a large-scale
empirical study in the Arcade Learning Environment (ALE) and, indeed, find
supporting evidence. We show that interference causes performance to plateau;
the network cannot train on segments beyond the plateau without degrading the
policy used to reach there. By synthetically controlling for interference, we
demonstrate performance boosts across architectures, learning algorithms and
environments. A more refined analysis shows that learning one segment of a game
often increases prediction errors elsewhere. Our study provides a clear
empirical link between catastrophic interference and sample efficiency in
reinforcement learning.
Related papers
- Multiple Descents in Unsupervised Learning: The Role of Noise, Domain Shift and Anomalies [14.399035468023161]
We study the presence of double descent in unsupervised learning, an area that has received little attention and is not yet fully understood.
We use synthetic and real data and identify model-wise, epoch-wise, and sample-wise double descent for various applications.
arXiv Detail & Related papers (2024-06-17T16:24:23Z) - Can Active Sampling Reduce Causal Confusion in Offline Reinforcement
Learning? [58.942118128503104]
Causal confusion is a phenomenon where an agent learns a policy that reflects imperfect spurious correlations in the data.
This phenomenon is particularly pronounced in domains such as robotics.
In this paper, we study causal confusion in offline reinforcement learning.
arXiv Detail & Related papers (2023-12-28T17:54:56Z) - On Continuity of Robust and Accurate Classifiers [3.8673630752805437]
It has been shown that adversarial training can improve the robustness of the hypothesis.
It has been suggested that robustness and accuracy of a hypothesis are at odds with each other.
In this paper, we put forth the alternative proposal that it is the continuity of a hypothesis that is incompatible with its robustness and accuracy.
arXiv Detail & Related papers (2023-09-29T08:14:25Z) - When are ensembles really effective? [49.37269057899679]
We study the question of when ensembling yields significant performance improvements in classification tasks.
We show that ensembling improves performance significantly whenever the disagreement rate is large relative to the average error rate.
We identify practical scenarios where ensembling does and does not result in large performance improvements.
arXiv Detail & Related papers (2023-05-21T01:36:25Z) - Chaos is a Ladder: A New Theoretical Understanding of Contrastive
Learning via Augmentation Overlap [64.60460828425502]
We propose a new guarantee on the downstream performance of contrastive learning.
Our new theory hinges on the insight that the support of different intra-class samples will become more overlapped under aggressive data augmentations.
We propose an unsupervised model selection metric ARC that aligns well with downstream accuracy.
arXiv Detail & Related papers (2022-03-25T05:36:26Z) - Where Did You Learn That From? Surprising Effectiveness of Membership
Inference Attacks Against Temporally Correlated Data in Deep Reinforcement
Learning [114.9857000195174]
A major challenge to widespread industrial adoption of deep reinforcement learning is the potential vulnerability to privacy breaches.
We propose an adversarial attack framework tailored for testing the vulnerability of deep reinforcement learning algorithms to membership inference attacks.
arXiv Detail & Related papers (2021-09-08T23:44:57Z) - Towards a practical measure of interference for reinforcement learning [37.1734757628306]
Catastrophic interference is common in many network-based learning systems.
We provide a definition of interference for control in reinforcement learning.
Our new interference measure allows us to ask novel scientific questions about commonly used deep learning architectures.
arXiv Detail & Related papers (2020-07-07T22:02:00Z) - Overfitting in adversarially robust deep learning [86.11788847990783]
We show that overfitting to the training set does in fact harm robust performance to a very large degree in adversarially robust training.
We also show that effects such as the double descent curve do still occur in adversarially trained models, yet fail to explain the observed overfitting.
arXiv Detail & Related papers (2020-02-26T15:40:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.