Adversarial Continual Learning
- URL: http://arxiv.org/abs/2003.09553v2
- Date: Tue, 21 Jul 2020 15:42:20 GMT
- Title: Adversarial Continual Learning
- Authors: Sayna Ebrahimi, Franziska Meier, Roberto Calandra, Trevor Darrell,
Marcus Rohrbach
- Abstract summary: We propose a hybrid continual learning framework that learns a disjoint representation for task-invariant and task-specific features.
Our model combines architecture growth to prevent forgetting of task-specific skills and an experience replay approach to preserve shared skills.
- Score: 99.56738010842301
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Continual learning aims to learn new tasks without forgetting previously
learned ones. We hypothesize that representations learned to solve each task in
a sequence have a shared structure while containing some task-specific
properties. We show that shared features are significantly less prone to
forgetting and propose a novel hybrid continual learning framework that learns
a disjoint representation for task-invariant and task-specific features
required to solve a sequence of tasks. Our model combines architecture growth
to prevent forgetting of task-specific skills and an experience replay approach
to preserve shared skills. We demonstrate our hybrid approach is effective in
avoiding forgetting and show it is superior to both architecture-based and
memory-based approaches on class incrementally learning of a single dataset as
well as a sequence of multiple datasets in image classification. Our code is
available at
\url{https://github.com/facebookresearch/Adversarial-Continual-Learning}.
Related papers
- Task agnostic continual learning with Pairwise layer architecture [0.0]
We show that we can improve the continual learning performance by replacing the final layer of our networks with our pairwise interaction layer.
The networks using this architecture show competitive performance in MNIST and FashionMNIST-based continual image classification experiments.
arXiv Detail & Related papers (2024-05-22T13:30:01Z) - Distribution Matching for Multi-Task Learning of Classification Tasks: a
Large-Scale Study on Faces & Beyond [62.406687088097605]
Multi-Task Learning (MTL) is a framework, where multiple related tasks are learned jointly and benefit from a shared representation space.
We show that MTL can be successful with classification tasks with little, or non-overlapping annotations.
We propose a novel approach, where knowledge exchange is enabled between the tasks via distribution matching.
arXiv Detail & Related papers (2024-01-02T14:18:11Z) - YOLOR-Based Multi-Task Learning [12.5920336941241]
Multi-task learning (MTL) aims to learn multiple tasks using a single model and jointly improve all of them assuming generalization and shared semantics.
We propose building on You Only Learn One Representation (YOLOR), a network architecture specifically designed for multitasking.
We find that our method achieves competitive performance on all tasks while maintaining a low parameter count and without any pre-training.
arXiv Detail & Related papers (2023-09-29T01:42:21Z) - Class-Incremental Learning via Knowledge Amalgamation [14.513858688486701]
Catastrophic forgetting has been a significant problem hindering the deployment of deep learning algorithms in the continual learning setting.
We put forward an alternative strategy to handle the catastrophic forgetting with knowledge amalgamation (CFA)
CFA learns a student network from multiple heterogeneous teacher models specializing in previous tasks and can be applied to current offline methods.
arXiv Detail & Related papers (2022-09-05T19:49:01Z) - Fast Inference and Transfer of Compositional Task Structures for
Few-shot Task Generalization [101.72755769194677]
We formulate it as a few-shot reinforcement learning problem where a task is characterized by a subtask graph.
Our multi-task subtask graph inferencer (MTSGI) first infers the common high-level task structure in terms of the subtask graph from the training tasks.
Our experiment results on 2D grid-world and complex web navigation domains show that the proposed method can learn and leverage the common underlying structure of the tasks for faster adaptation to the unseen tasks.
arXiv Detail & Related papers (2022-05-25T10:44:25Z) - Combining Modular Skills in Multitask Learning [149.8001096811708]
A modular design encourages neural models to disentangle and recombine different facets of knowledge to generalise more systematically to new tasks.
In this work, we assume each task is associated with a subset of latent discrete skills from a (potentially small) inventory.
We find that the modular design of a network significantly increases sample efficiency in reinforcement learning and few-shot generalisation in supervised learning.
arXiv Detail & Related papers (2022-02-28T16:07:19Z) - Encoders and Ensembles for Task-Free Continual Learning [15.831773437720429]
We present an architecture that is effective for continual learning in an especially demanding setting, where task boundaries do not exist or are unknown.
We show that models trained with the architecture are state-of-the-art for the task-free setting on standard image classification continual learning benchmarks.
We also show that the architecture learns well in a fully incremental setting, where one class is learned at a time, and we demonstrate its effectiveness in this setting with up to 100 classes.
arXiv Detail & Related papers (2021-05-27T17:34:31Z) - Continual Learning in Low-rank Orthogonal Subspaces [86.36417214618575]
In continual learning (CL), a learner is faced with a sequence of tasks, arriving one after the other, and the goal is to remember all the tasks once the learning experience is finished.
The prior art in CL uses episodic memory, parameter regularization or network structures to reduce interference among tasks, but in the end, all the approaches learn different tasks in a joint vector space.
We propose to learn tasks in different (low-rank) vector subspaces that are kept orthogonal to each other in order to minimize interference.
arXiv Detail & Related papers (2020-10-22T12:07:43Z) - Task-Feature Collaborative Learning with Application to Personalized
Attribute Prediction [166.87111665908333]
We propose a novel multi-task learning method called Task-Feature Collaborative Learning (TFCL)
Specifically, we first propose a base model with a heterogeneous block-diagonal structure regularizer to leverage the collaborative grouping of features and tasks.
As a practical extension, we extend the base model by allowing overlapping features and differentiating the hard tasks.
arXiv Detail & Related papers (2020-04-29T02:32:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.