Incremental Learning of Structured Memory via Closed-Loop Transcription
- URL: http://arxiv.org/abs/2202.05411v3
- Date: Wed, 7 Jun 2023 05:00:32 GMT
- Title: Incremental Learning of Structured Memory via Closed-Loop Transcription
- Authors: Shengbang Tong, Xili Dai, Ziyang Wu, Mingyang Li, Brent Yi, Yi Ma
- Abstract summary: This work proposes a minimal computational model for learning structured memories of multiple object classes in an incremental setting.
Our method is simpler than existing approaches for incremental learning, and more efficient in terms of model size, storage, and computation.
Experimental results show that our method can effectively alleviate catastrophic forgetting, achieving significantly better performance than prior work of generative replay.
- Score: 20.255633973040183
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This work proposes a minimal computational model for learning structured
memories of multiple object classes in an incremental setting. Our approach is
based on establishing a closed-loop transcription between the classes and a
corresponding set of subspaces, known as a linear discriminative
representation, in a low-dimensional feature space. Our method is simpler than
existing approaches for incremental learning, and more efficient in terms of
model size, storage, and computation: it requires only a single, fixed-capacity
autoencoding network with a feature space that is used for both discriminative
and generative purposes. Network parameters are optimized simultaneously
without architectural manipulations, by solving a constrained minimax game
between the encoding and decoding maps over a single rate reduction-based
objective. Experimental results show that our method can effectively alleviate
catastrophic forgetting, achieving significantly better performance than prior
work of generative replay on MNIST, CIFAR-10, and ImageNet-50, despite
requiring fewer resources. Source code can be found at
https://github.com/tsb0601/i-CTRL
Related papers
- SHERL: Synthesizing High Accuracy and Efficient Memory for Resource-Limited Transfer Learning [63.93193829913252]
We propose an innovative METL strategy called SHERL for resource-limited scenarios.
In the early route, intermediate outputs are consolidated via an anti-redundancy operation.
In the late route, utilizing minimal late pre-trained layers could alleviate the peak demand on memory overhead.
arXiv Detail & Related papers (2024-07-10T10:22:35Z) - VeLoRA: Memory Efficient Training using Rank-1 Sub-Token Projections [35.133698935322634]
Large language models (LLMs) have recently emerged as powerful tools for tackling many language-processing tasks.
We identify and characterise the important components needed for effective model convergence using gradient descent.
This result leads us to a cheap and memory-efficient algorithm for both fine-tuning and pre-training LLMs.
arXiv Detail & Related papers (2024-05-28T09:23:14Z) - Hierarchical Context Merging: Better Long Context Understanding for Pre-trained LLMs [61.40047491337793]
We present Hierarchical cOntext MERging (HOMER), a new training-free scheme designed to overcome the limitations of large language models.
HomeR uses a divide-and-conquer algorithm, dividing long inputs into manageable chunks.
A token reduction technique precedes each merging, ensuring memory usage efficiency.
arXiv Detail & Related papers (2024-04-16T06:34:08Z) - Deep Unlearning: Fast and Efficient Gradient-free Approach to Class Forgetting [9.91998873101083]
We introduce a novel class unlearning algorithm designed to strategically eliminate specific classes from the learned model.
Our algorithm exhibits competitive unlearning performance and resilience against Membership Inference Attacks (MIA)
arXiv Detail & Related papers (2023-12-01T18:29:08Z) - MultIOD: Rehearsal-free Multihead Incremental Object Detector [17.236182938227163]
We propose MultIOD, a class-incremental object detector based on CenterNet.
We employ transfer learning between classes learned initially and those learned incrementally to tackle catastrophic forgetting.
Results show that our method outperforms state-of-the-art methods on two Pascal VOC datasets.
arXiv Detail & Related papers (2023-09-11T09:32:45Z) - Just One Byte (per gradient): A Note on Low-Bandwidth Decentralized
Language Model Finetuning Using Shared Randomness [86.61582747039053]
Language model training in distributed settings is limited by the communication cost of exchanges.
We extend recent work using shared randomness to perform distributed fine-tuning with low bandwidth.
arXiv Detail & Related papers (2023-06-16T17:59:51Z) - A Model or 603 Exemplars: Towards Memory-Efficient Class-Incremental
Learning [56.450090618578]
Class-Incremental Learning (CIL) aims to train a model with limited memory size to meet this requirement.
We show that when counting the model size into the total budget and comparing methods with aligned memory size, saving models do not consistently work.
We propose a simple yet effective baseline, denoted as MEMO for Memory-efficient Expandable MOdel.
arXiv Detail & Related papers (2022-05-26T08:24:01Z) - Decoupled and Memory-Reinforced Networks: Towards Effective Feature
Learning for One-Step Person Search [65.51181219410763]
One-step methods have been developed to handle pedestrian detection and identification sub-tasks using a single network.
There are two major challenges in the current one-step approaches.
We propose a decoupled and memory-reinforced network (DMRNet) to overcome these problems.
arXiv Detail & Related papers (2021-02-22T06:19:45Z) - Fast Few-Shot Classification by Few-Iteration Meta-Learning [173.32497326674775]
We introduce a fast optimization-based meta-learning method for few-shot classification.
Our strategy enables important aspects of the base learner objective to be learned during meta-training.
We perform a comprehensive experimental analysis, demonstrating the speed and effectiveness of our approach.
arXiv Detail & Related papers (2020-10-01T15:59:31Z) - SpaceNet: Make Free Space For Continual Learning [15.914199054779438]
We propose a novel architectural-based method referred as SpaceNet for class incremental learning scenario.
SpaceNet trains sparse deep neural networks from scratch in an adaptive way that compresses the sparse connections of each task in a compact number of neurons.
Experimental results show the robustness of our proposed method against catastrophic forgetting old tasks and the efficiency of SpaceNet in utilizing the available capacity of the model.
arXiv Detail & Related papers (2020-07-15T11:21:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.