Negotiated Representations to Prevent Forgetting in Machine Learning
Applications
- URL: http://arxiv.org/abs/2312.00237v1
- Date: Thu, 30 Nov 2023 22:43:50 GMT
- Title: Negotiated Representations to Prevent Forgetting in Machine Learning
Applications
- Authors: Nuri Korhan, Ceren \"Oner
- Abstract summary: Catastrophic forgetting is a significant challenge in the field of machine learning.
We propose a novel method for preventing catastrophic forgetting in machine learning applications.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Catastrophic forgetting is a significant challenge in the field of machine
learning, particularly in neural networks. When a neural network learns to
perform well on a new task, it often forgets its previously acquired knowledge
or experiences. This phenomenon occurs because the network adjusts its weights
and connections to minimize the loss on the new task, which can inadvertently
overwrite or disrupt the representations that were crucial for the previous
tasks. As a result, the the performance of the network on earlier tasks
deteriorates, limiting its ability to learn and adapt to a sequence of tasks.
In this paper, we propose a novel method for preventing catastrophic forgetting
in machine learning applications, specifically focusing on neural networks. Our
approach aims to preserve the knowledge of the network across multiple tasks
while still allowing it to learn new information effectively. We demonstrate
the effectiveness of our method by conducting experiments on various benchmark
datasets, including Split MNIST, Split CIFAR10, Split Fashion MNIST, and Split
CIFAR100. These datasets are created by dividing the original datasets into
separate, non overlapping tasks, simulating a continual learning scenario where
the model needs to learn multiple tasks sequentially without forgetting the
previous ones. Our proposed method tackles the catastrophic forgetting problem
by incorporating negotiated representations into the learning process, which
allows the model to maintain a balance between retaining past experiences and
adapting to new tasks. By evaluating our method on these challenging datasets,
we aim to showcase its potential for addressing catastrophic forgetting and
improving the performance of neural networks in continual learning settings.
Related papers
- Towards Scalable and Versatile Weight Space Learning [51.78426981947659]
This paper introduces the SANE approach to weight-space learning.
Our method extends the idea of hyper-representations towards sequential processing of subsets of neural network weights.
arXiv Detail & Related papers (2024-06-14T13:12:07Z) - Look-Ahead Selective Plasticity for Continual Learning of Visual Tasks [9.82510084910641]
We propose a new mechanism that takes place during task boundaries, i.e., when one task finishes and another starts.
We evaluate the proposed methods on benchmark computer vision datasets including CIFAR10 and TinyImagenet.
arXiv Detail & Related papers (2023-11-02T22:00:23Z) - Provable Multi-Task Representation Learning by Two-Layer ReLU Neural Networks [69.38572074372392]
We present the first results proving that feature learning occurs during training with a nonlinear model on multiple tasks.
Our key insight is that multi-task pretraining induces a pseudo-contrastive loss that favors representations that align points that typically have the same label across tasks.
arXiv Detail & Related papers (2023-07-13T16:39:08Z) - Complementary Learning Subnetworks for Parameter-Efficient
Class-Incremental Learning [40.13416912075668]
We propose a rehearsal-free CIL approach that learns continually via the synergy between two Complementary Learning Subnetworks.
Our method achieves competitive results against state-of-the-art methods, especially in accuracy gain, memory cost, training efficiency, and task-order.
arXiv Detail & Related papers (2023-06-21T01:43:25Z) - IF2Net: Innately Forgetting-Free Networks for Continual Learning [49.57495829364827]
Continual learning can incrementally absorb new concepts without interfering with previously learned knowledge.
Motivated by the characteristics of neural networks, we investigated how to design an Innately Forgetting-Free Network (IF2Net)
IF2Net allows a single network to inherently learn unlimited mapping rules without telling task identities at test time.
arXiv Detail & Related papers (2023-06-18T05:26:49Z) - Learning Bayesian Sparse Networks with Full Experience Replay for
Continual Learning [54.7584721943286]
Continual Learning (CL) methods aim to enable machine learning models to learn new tasks without catastrophic forgetting of those that have been previously mastered.
Existing CL approaches often keep a buffer of previously-seen samples, perform knowledge distillation, or use regularization techniques towards this goal.
We propose to only activate and select sparse neurons for learning current and past tasks at any stage.
arXiv Detail & Related papers (2022-02-21T13:25:03Z) - Center Loss Regularization for Continual Learning [0.0]
In general, neural networks lack the ability to learn different tasks sequentially.
Our approach remembers old tasks by projecting the representations of new tasks close to that of old tasks.
We demonstrate that our approach is scalable, effective, and gives competitive performance compared to state-of-the-art continual learning methods.
arXiv Detail & Related papers (2021-10-21T17:46:44Z) - Incremental Embedding Learning via Zero-Shot Translation [65.94349068508863]
Current state-of-the-art incremental learning methods tackle catastrophic forgetting problem in traditional classification networks.
We propose a novel class-incremental method for embedding network, named as zero-shot translation class-incremental method (ZSTCI)
In addition, ZSTCI can easily be combined with existing regularization-based incremental learning methods to further improve performance of embedding networks.
arXiv Detail & Related papers (2020-12-31T08:21:37Z) - Semantic Drift Compensation for Class-Incremental Learning [48.749630494026086]
Class-incremental learning of deep networks sequentially increases the number of classes to be classified.
We propose a new method to estimate the drift, called semantic drift, of features and compensate for it without the need of any exemplars.
arXiv Detail & Related papers (2020-04-01T13:31:19Z) - Adversarial Incremental Learning [0.0]
Deep learning can forget previously learned information upon learning new tasks where previous data is not available.
We propose an adversarial discriminator based method that does not make use of old data at all while training on new tasks.
We are able to outperform other state-of-the-art methods on CIFAR-100, SVHN, and MNIST datasets.
arXiv Detail & Related papers (2020-01-30T02:25:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.