Remembering for the Right Reasons: Explanations Reduce Catastrophic
Forgetting
- URL: http://arxiv.org/abs/2010.01528v2
- Date: Mon, 3 May 2021 03:26:30 GMT
- Title: Remembering for the Right Reasons: Explanations Reduce Catastrophic
Forgetting
- Authors: Sayna Ebrahimi, Suzanne Petryk, Akash Gokul, William Gan, Joseph E.
Gonzalez, Marcus Rohrbach, Trevor Darrell
- Abstract summary: We propose a novel training paradigm called Remembering for the Right Reasons (RRR)
RRR stores visual model explanations for each example in the buffer and ensures the model has "the right reasons" for its predictions.
We demonstrate how RRR can be easily added to any memory or regularization-based approach and results in reduced forgetting.
- Score: 100.75479161884935
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The goal of continual learning (CL) is to learn a sequence of tasks without
suffering from the phenomenon of catastrophic forgetting. Previous work has
shown that leveraging memory in the form of a replay buffer can reduce
performance degradation on prior tasks. We hypothesize that forgetting can be
further reduced when the model is encouraged to remember the \textit{evidence}
for previously made decisions. As a first step towards exploring this
hypothesis, we propose a simple novel training paradigm, called Remembering for
the Right Reasons (RRR), that additionally stores visual model explanations for
each example in the buffer and ensures the model has "the right reasons" for
its predictions by encouraging its explanations to remain consistent with those
used to make decisions at training time. Without this constraint, there is a
drift in explanations and increase in forgetting as conventional continual
learning algorithms learn new tasks. We demonstrate how RRR can be easily added
to any memory or regularization-based approach and results in reduced
forgetting, and more importantly, improved model explanations. We have
evaluated our approach in the standard and few-shot settings and observed a
consistent improvement across various CL approaches using different
architectures and techniques to generate model explanations and demonstrated
our approach showing a promising connection between explainability and
continual learning. Our code is available at
\url{https://github.com/SaynaEbrahimi/Remembering-for-the-Right-Reasons}.
Related papers
- Temporal-Difference Variational Continual Learning [89.32940051152782]
A crucial capability of Machine Learning models in real-world applications is the ability to continuously learn new tasks.
In Continual Learning settings, models often struggle to balance learning new tasks with retaining previous knowledge.
We propose new learning objectives that integrate the regularization effects of multiple previous posterior estimations.
arXiv Detail & Related papers (2024-10-10T10:58:41Z) - Reason for Future, Act for Now: A Principled Framework for Autonomous
LLM Agents with Provable Sample Efficiency [53.8779374188643]
We propose a principled framework with provable regret guarantees to orchestrate reasoning and acting.
Specifically, we design a prompt template for reasoning that learns from the memory buffer and plans a future trajectory over a long horizon.
At each step, the LLM agent takes the initial action of the planned trajectory ("act for now"), stores the collected feedback in the memory buffer, and reinvokes the reasoning routine to replan the future trajectory from the new state.
arXiv Detail & Related papers (2023-09-29T16:36:39Z) - Mitigating Catastrophic Forgetting in Task-Incremental Continual
Learning with Adaptive Classification Criterion [50.03041373044267]
We propose a Supervised Contrastive learning framework with adaptive classification criterion for Continual Learning.
Experiments show that CFL achieves state-of-the-art performance and has a stronger ability to overcome compared with the classification baselines.
arXiv Detail & Related papers (2023-05-20T19:22:40Z) - Learning with Explanation Constraints [91.23736536228485]
We provide a learning theoretic framework to analyze how explanations can improve the learning of our models.
We demonstrate the benefits of our approach over a large array of synthetic and real-world experiments.
arXiv Detail & Related papers (2023-03-25T15:06:47Z) - Harnessing the Power of Explanations for Incremental Training: A
LIME-Based Approach [6.244905619201076]
In this work, model explanations are fed back to the feed-forward training to help the model generalize better.
The framework incorporates the custom weighted loss with Elastic Weight Consolidation (EWC) to maintain performance in sequential testing sets.
The proposed custom training procedure results in a consistent enhancement of accuracy ranging from 0.5% to 1.5% throughout all phases of the incremental learning setup.
arXiv Detail & Related papers (2022-11-02T18:16:17Z) - Learning an evolved mixture model for task-free continual learning [11.540150938141034]
We address the Task-Free Continual Learning (TFCL) in which a model is trained on non-stationary data streams with no explicit task information.
We introduce two simple dropout mechanisms to selectively remove stored examples in order to avoid memory overload.
arXiv Detail & Related papers (2022-07-11T16:01:27Z) - Decoupling Knowledge from Memorization: Retrieval-augmented Prompt
Learning [113.58691755215663]
We develop RetroPrompt to help a model strike a balance between generalization and memorization.
In contrast with vanilla prompt learning, RetroPrompt constructs an open-book knowledge-store from training instances.
Extensive experiments demonstrate that RetroPrompt can obtain better performance in both few-shot and zero-shot settings.
arXiv Detail & Related papers (2022-05-29T16:07:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.