Multiple Modes for Continual Learning
- URL: http://arxiv.org/abs/2209.14996v1
- Date: Thu, 29 Sep 2022 17:55:32 GMT
- Title: Multiple Modes for Continual Learning
- Authors: Siddhartha Datta, Nigel Shadbolt
- Abstract summary: Adapting model parameters to incoming streams of data is a crucial factor to deep learning scalability.
We formulate a trade-off between constructing multiple parameter modes and allocating tasks per mode.
We empirically demonstrate improvements over baseline continual learning strategies.
- Score: 8.782809316491948
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Adapting model parameters to incoming streams of data is a crucial factor to
deep learning scalability. Interestingly, prior continual learning strategies
in online settings inadvertently anchor their updated parameters to a local
parameter subspace to remember old tasks, else drift away from the subspace and
forget. From this observation, we formulate a trade-off between constructing
multiple parameter modes and allocating tasks per mode. Mode-Optimized Task
Allocation (MOTA), our contributed adaptation strategy, trains multiple modes
in parallel, then optimizes task allocation per mode. We empirically
demonstrate improvements over baseline continual learning strategies and across
varying distribution shifts, namely sub-population, domain, and task shift.
Related papers
- Task-Aware Harmony Multi-Task Decision Transformer for Offline Reinforcement Learning [70.96345405979179]
The purpose of offline multi-task reinforcement learning (MTRL) is to develop a unified policy applicable to diverse tasks without the need for online environmental interaction.
variations in task content and complexity pose significant challenges in policy formulation.
We introduce the Harmony Multi-Task Decision Transformer (HarmoDT), a novel solution designed to identify an optimal harmony subspace of parameters for each task.
arXiv Detail & Related papers (2024-11-02T05:49:14Z) - HarmoDT: Harmony Multi-Task Decision Transformer for Offline Reinforcement Learning [72.25707314772254]
We introduce the Harmony Multi-Task Decision Transformer (HarmoDT), a novel solution designed to identify an optimal harmony subspace of parameters for each task.
The upper level of this framework is dedicated to learning a task-specific mask that delineates the harmony subspace, while the inner level focuses on updating parameters to enhance the overall performance of the unified policy.
arXiv Detail & Related papers (2024-05-28T11:41:41Z) - MetaModulation: Learning Variational Feature Hierarchies for Few-Shot
Learning with Fewer Tasks [63.016244188951696]
We propose a method for few-shot learning with fewer tasks, which is by metaulation.
We modify parameters at various batch levels to increase the meta-training tasks.
We also introduce learning variational feature hierarchies by incorporating the variationalulation.
arXiv Detail & Related papers (2023-05-17T15:47:47Z) - Exploiting Style Transfer-based Task Augmentation for Cross-Domain
Few-Shot Learning [4.678020383205135]
In cross-domain few-shot learning, the model trained on source domains struggles to generalize to the target domain.
We propose Task Augmented Meta-Learning (TAML) to conduct style transfer-based task augmentation.
The proposed TAML increases the diversity of styles of training tasks, and contributes to training a model with better domain generalization ability.
arXiv Detail & Related papers (2023-01-19T07:32:23Z) - Improving Meta-learning for Low-resource Text Classification and
Generation via Memory Imitation [87.98063273826702]
We propose a memory imitation meta-learning (MemIML) method that enhances the model's reliance on support sets for task adaptation.
A theoretical analysis is provided to prove the effectiveness of our method.
arXiv Detail & Related papers (2022-03-22T12:41:55Z) - Efficient Feature Transformations for Discriminative and Generative
Continual Learning [98.10425163678082]
We propose a simple task-specific feature map transformation strategy for continual learning.
Theses provide powerful flexibility for learning new tasks, achieved with minimal parameters added to the base architecture.
We demonstrate the efficacy and efficiency of our method with an extensive set of experiments in discriminative (CIFAR-100 and ImageNet-1K) and generative sequences of tasks.
arXiv Detail & Related papers (2021-03-25T01:48:14Z) - Efficient Continual Adaptation for Generative Adversarial Networks [97.20244383723853]
We present a continual learning approach for generative adversarial networks (GANs)
Our approach is based on learning a set of global and task-specific parameters.
We show that the feature-map transformation based approach outperforms state-of-the-art continual GANs methods.
arXiv Detail & Related papers (2021-03-06T05:09:37Z) - Neuromodulated Neural Architectures with Local Error Signals for
Memory-Constrained Online Continual Learning [4.2903672492917755]
We develop a biologically-inspired light weight neural network architecture that incorporates local learning and neuromodulation.
We demonstrate the efficacy of our approach on both single task and continual learning setting.
arXiv Detail & Related papers (2020-07-16T07:41:23Z) - Multi-Task Reinforcement Learning with Soft Modularization [25.724764855681137]
Multi-task learning is a very challenging problem in reinforcement learning.
We introduce an explicit modularization technique on policy representation to alleviate this optimization issue.
We show our method improves both sample efficiency and performance over strong baselines by a large margin.
arXiv Detail & Related papers (2020-03-30T17:47:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.