Memory Bounds for Continual Learning
- URL: http://arxiv.org/abs/2204.10830v1
- Date: Fri, 22 Apr 2022 17:19:50 GMT
- Title: Memory Bounds for Continual Learning
- Authors: Xi Chen, Christos Papadimitriou and Binghui Peng
- Abstract summary: Continual learning, or lifelong learning, is a formidable current challenge to machine learning.
We make novel uses of communication complexity to establish that any continual learner, even an improper one, needs memory that grows linearly with $k$.
- Score: 13.734474418577188
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Continual learning, or lifelong learning, is a formidable current challenge
to machine learning. It requires the learner to solve a sequence of $k$
different learning tasks, one after the other, while retaining its aptitude for
earlier tasks; the continual learner should scale better than the obvious
solution of developing and maintaining a separate learner for each of the $k$
tasks. We embark on a complexity-theoretic study of continual learning in the
PAC framework. We make novel uses of communication complexity to establish that
any continual learner, even an improper one, needs memory that grows linearly
with $k$, strongly suggesting that the problem is intractable. When
logarithmically many passes over the learning tasks are allowed, we provide an
algorithm based on multiplicative weights update whose memory requirement
scales well; we also establish that improper learning is necessary for such
performance. We conjecture that these results may lead to new promising
approaches to continual learning.
Related papers
- Continual Learning of Numerous Tasks from Long-tail Distributions [17.706669222987273]
Continual learning focuses on developing models that learn and adapt to new tasks while retaining previously acquired knowledge.
Existing continual learning algorithms usually involve a small number of tasks with uniform sizes and may not accurately represent real-world learning scenarios.
We propose a method that reuses the states in Adam by maintaining a weighted average of the second moments from previous tasks.
We demonstrate that our method, compatible with most existing continual learning algorithms, effectively reduces forgetting with only a small amount of additional computational or memory costs.
arXiv Detail & Related papers (2024-04-03T13:56:33Z) - Multitask Learning with No Regret: from Improved Confidence Bounds to
Active Learning [79.07658065326592]
Quantifying uncertainty in the estimated tasks is of pivotal importance for many downstream applications, such as online or active learning.
We provide novel multitask confidence intervals in the challenging setting when neither the similarity between tasks nor the tasks' features are available to the learner.
We propose a novel online learning algorithm that achieves such improved regret without knowing this parameter in advance.
arXiv Detail & Related papers (2023-08-03T13:08:09Z) - How Efficient Are Today's Continual Learning Algorithms? [31.120016345185217]
Supervised continual learning involves updating a deep neural network (DNN) from an ever-growing stream of labeled data.
One of the major motivations behind continual learning is being able to efficiently update a network with new information, rather than retraining from scratch on the training dataset as it grows over time.
Here, we study recent methods for incremental class learning and illustrate that many are highly inefficient in terms of compute, memory, and storage.
arXiv Detail & Related papers (2023-03-29T18:52:10Z) - Teacher-student curriculum learning for reinforcement learning [1.7259824817932292]
Reinforcement learning (rl) is a popular paradigm for sequential decision making problems.
The sample inefficiency of deep reinforcement learning methods is a significant obstacle when applying rl to real-world problems.
We propose a teacher-student curriculum learning setting where we simultaneously train a teacher that selects tasks for the student while the student learns how to solve the selected task.
arXiv Detail & Related papers (2022-10-31T14:45:39Z) - Reset-Free Reinforcement Learning via Multi-Task Learning: Learning
Dexterous Manipulation Behaviors without Human Intervention [67.1936055742498]
We show that multi-task learning can effectively scale reset-free learning schemes to much more complex problems.
This work shows the ability to learn dexterous manipulation behaviors in the real world with RL without any human intervention.
arXiv Detail & Related papers (2021-04-22T17:38:27Z) - Variable-Shot Adaptation for Online Meta-Learning [123.47725004094472]
We study the problem of learning new tasks from a small, fixed number of examples, by meta-learning across static data from a set of previous tasks.
We find that meta-learning solves the full task set with fewer overall labels and greater cumulative performance, compared to standard supervised methods.
These results suggest that meta-learning is an important ingredient for building learning systems that continuously learn and improve over a sequence of problems.
arXiv Detail & Related papers (2020-12-14T18:05:24Z) - Continual Learning in Low-rank Orthogonal Subspaces [86.36417214618575]
In continual learning (CL), a learner is faced with a sequence of tasks, arriving one after the other, and the goal is to remember all the tasks once the learning experience is finished.
The prior art in CL uses episodic memory, parameter regularization or network structures to reduce interference among tasks, but in the end, all the approaches learn different tasks in a joint vector space.
We propose to learn tasks in different (low-rank) vector subspaces that are kept orthogonal to each other in order to minimize interference.
arXiv Detail & Related papers (2020-10-22T12:07:43Z) - Bilevel Continual Learning [76.50127663309604]
We present a novel framework of continual learning named "Bilevel Continual Learning" (BCL)
Our experiments on continual learning benchmarks demonstrate the efficacy of the proposed BCL compared to many state-of-the-art methods.
arXiv Detail & Related papers (2020-07-30T16:00:23Z) - Curriculum Learning for Reinforcement Learning Domains: A Framework and
Survey [53.73359052511171]
Reinforcement learning (RL) is a popular paradigm for addressing sequential decision tasks in which the agent has only limited environmental feedback.
We present a framework for curriculum learning (CL) in RL, and use it to survey and classify existing CL methods in terms of their assumptions, capabilities, and goals.
arXiv Detail & Related papers (2020-03-10T20:41:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.