RMM: Reinforced Memory Management for Class-Incremental Learning
- URL: http://arxiv.org/abs/2301.05792v1
- Date: Sat, 14 Jan 2023 00:07:47 GMT
- Title: RMM: Reinforced Memory Management for Class-Incremental Learning
- Authors: Yaoyao Liu, Bernt Schiele, Qianru Sun
- Abstract summary: Class-Incremental Learning (CIL) trains classifiers under a strict memory budget.
Existing methods use a static and ad hoc strategy for memory allocation, which is often sub-optimal.
We propose a dynamic memory management strategy that is optimized for the incremental phases and different object classes.
- Score: 102.20140790771265
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Class-Incremental Learning (CIL) [40] trains classifiers under a strict
memory budget: in each incremental phase, learning is done for new data, most
of which is abandoned to free space for the next phase. The preserved data are
exemplars used for replaying. However, existing methods use a static and ad hoc
strategy for memory allocation, which is often sub-optimal. In this work, we
propose a dynamic memory management strategy that is optimized for the
incremental phases and different object classes. We call our method reinforced
memory management (RMM), leveraging reinforcement learning. RMM training is not
naturally compatible with CIL as the past, and future data are strictly
non-accessible during the incremental phases. We solve this by training the
policy function of RMM on pseudo CIL tasks, e.g., the tasks built on the data
of the 0-th phase, and then applying it to target tasks. RMM propagates two
levels of actions: Level-1 determines how to split the memory between old and
new classes, and Level-2 allocates memory for each specific class. In essence,
it is an optimizable and general method for memory management that can be used
in any replaying-based CIL method. For evaluation, we plug RMM into two
top-performing baselines (LUCIR+AANets and POD+AANets [30]) and conduct
experiments on three benchmarks (CIFAR-100, ImageNet-Subset, and
ImageNet-Full). Our results show clear improvements, e.g., boosting POD+AANets
by 3.6%, 4.4%, and 1.9% in the 25-Phase settings of the above benchmarks,
respectively.
Related papers
- vTensor: Flexible Virtual Tensor Management for Efficient LLM Serving [53.972175896814505]
Large Language Models (LLMs) are widely used across various domains, processing millions of daily requests.
Large Language Models (LLMs) are widely used across various domains, processing millions of daily requests.
arXiv Detail & Related papers (2024-07-22T14:37:58Z) - Efficiently Training 7B LLM with 1 Million Sequence Length on 8 GPUs [24.066283519769968]
Large Language Models (LLMs) have been trained using extended context lengths to foster more creative applications.
We propose MEMO, a novel framework for fine-grained activation memory management.
We show that MEMO achieves an average of 2.42x and 2.26x MFU compared to Megatron-LM and DeepSpeed.
arXiv Detail & Related papers (2024-07-16T18:59:49Z) - PARMESAN: Parameter-Free Memory Search and Transduction for Dense Prediction Tasks [5.5127111704068374]
This work addresses flexibility in deep learning by means of transductive reasoning.
We propose PARMESAN, a scalable method which leverages a memory module for solving dense prediction tasks.
Our method is compatible with commonly used architectures and canonically transfers to 1D, 2D, and 3D grid-based data.
arXiv Detail & Related papers (2024-03-18T12:55:40Z) - GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection [133.45193150403537]
Training Large Language Models (LLMs) presents significant memory challenges due to the growing size of weights and GPU states.
In this work, we propose Gradient Low-Rank Projection (GaLore) as a memory-efficient training strategy.
Our 8-bit GaLore further reduces memory by up to 82.5% and total training memory by 63.3%, compared to a BF16 baseline.
arXiv Detail & Related papers (2024-03-06T07:29:57Z) - MAMBA: Multi-level Aggregation via Memory Bank for Video Object
Detection [35.16197118579414]
We propose a multi-level aggregation architecture via memory bank called MAMBA.
Specifically, our memory bank employs two novel operations to eliminate the disadvantages of existing methods.
Compared with existing state-of-the-art methods, our method achieves superior performance in terms of both speed and accuracy.
arXiv Detail & Related papers (2024-01-18T12:13:06Z) - Continual Referring Expression Comprehension via Dual Modular
Memorization [133.46886428655426]
Referring Expression (REC) aims to localize an image region of a given object described by a natural-language expression.
Existing REC algorithms make a strong assumption that training data feeding into a model are given upfront, which degrades its practicality for real-world scenarios.
In this paper, we propose Continual Referring Expression (CREC), a new setting for REC, where a model is learning on a stream of incoming tasks.
In order to continuously improve the model on sequential tasks without forgetting prior learned knowledge and without repeatedly re-training from a scratch, we propose an effective baseline method named Dual Modular Memorization
arXiv Detail & Related papers (2023-11-25T02:58:51Z) - A Model or 603 Exemplars: Towards Memory-Efficient Class-Incremental
Learning [56.450090618578]
Class-Incremental Learning (CIL) aims to train a model with limited memory size to meet this requirement.
We show that when counting the model size into the total budget and comparing methods with aligned memory size, saving models do not consistently work.
We propose a simple yet effective baseline, denoted as MEMO for Memory-efficient Expandable MOdel.
arXiv Detail & Related papers (2022-05-26T08:24:01Z) - Memory-Guided Semantic Learning Network for Temporal Sentence Grounding [55.31041933103645]
We propose a memory-augmented network that learns and memorizes the rarely appeared content in TSG tasks.
MGSL-Net consists of three main parts: a cross-modal inter-action module, a memory augmentation module, and a heterogeneous attention module.
arXiv Detail & Related papers (2022-01-03T02:32:06Z) - The act of remembering: a study in partially observable reinforcement
learning [24.945756871291348]
Reinforcement Learning (RL) agents typically learn memoryless policies that only consider the last observation when selecting actions.
We provide the agent with an external memory and additional actions to control what, if anything, is written to the memory.
Our novel forms of memory outperform binary and LSTM-based memory in well-established partially observable domains.
arXiv Detail & Related papers (2020-10-05T02:56:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.