From MNIST to ImageNet and Back: Benchmarking Continual Curriculum
Learning
- URL: http://arxiv.org/abs/2303.11076v1
- Date: Thu, 16 Mar 2023 18:11:19 GMT
- Title: From MNIST to ImageNet and Back: Benchmarking Continual Curriculum
Learning
- Authors: Kamil Faber, Dominik Zurek, Marcin Pietron, Nathalie Japkowicz,
Antonio Vergari, Roberto Corizzo
- Abstract summary: Continual learning (CL) is one of the most promising trends in machine learning research.
We introduce two novel CL benchmarks that involve multiple heterogeneous tasks from six image datasets.
We additionally structure our benchmarks so that tasks are presented in increasing and decreasing order of complexity.
- Score: 9.104068727716294
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Continual learning (CL) is one of the most promising trends in recent machine
learning research. Its goal is to go beyond classical assumptions in machine
learning and develop models and learning strategies that present high
robustness in dynamic environments. The landscape of CL research is fragmented
into several learning evaluation protocols, comprising different learning
tasks, datasets, and evaluation metrics. Additionally, the benchmarks adopted
so far are still distant from the complexity of real-world scenarios, and are
usually tailored to highlight capabilities specific to certain strategies. In
such a landscape, it is hard to objectively assess strategies. In this work, we
fill this gap for CL on image data by introducing two novel CL benchmarks that
involve multiple heterogeneous tasks from six image datasets, with varying
levels of complexity and quality. Our aim is to fairly evaluate current
state-of-the-art CL strategies on a common ground that is closer to complex
real-world scenarios. We additionally structure our benchmarks so that tasks
are presented in increasing and decreasing order of complexity -- according to
a curriculum -- in order to evaluate if current CL models are able to exploit
structure across tasks. We devote particular emphasis to providing the CL
community with a rigorous and reproducible evaluation protocol for measuring
the ability of a model to generalize and not to forget while learning.
Furthermore, we provide an extensive experimental evaluation showing that
popular CL strategies, when challenged with our benchmarks, yield sub-par
performance, high levels of forgetting, and present a limited ability to
effectively leverage curriculum task ordering. We believe that these results
highlight the need for rigorous comparisons in future CL works as well as pave
the way to design new CL strategies that are able to deal with more complex
scenarios.
Related papers
- Does the Definition of Difficulty Matter? Scoring Functions and their Role for Curriculum Learning [42.4526628515253]
Curriculum learning (CL) describes a machine learning training strategy in which samples are gradually introduced into the training process based on their difficulty.
We study the robustness and similarity of the most common scoring functions for sample difficulty estimation.
We find that the robustness of scoring functions across random seeds positively correlates with CL performance.
arXiv Detail & Related papers (2024-11-01T18:55:31Z) - Realistic Continual Learning Approach using Pre-trained Models [1.2582887633807602]
We introduce Realistic Continual Learning (RealCL), a novel CL paradigm where class distributions across tasks are random.
We also present CLARE (Continual Learning Approach with pRE-trained models for RealCL scenarios), a pre-trained model-based solution designed to integrate new knowledge while preserving past learning.
arXiv Detail & Related papers (2024-04-11T13:19:46Z) - Hyperparameters in Continual Learning: A Reality Check [53.30082523545212]
Continual learning (CL) aims to train a model on a sequence of tasks while balancing the trade-off between plasticity (learning new tasks) and stability (retaining prior knowledge)
The dominantly adopted conventional evaluation protocol for CL algorithms selects the best hyper parameters in a given scenario and then evaluates the algorithms in the same scenario.
This protocol has significant shortcomings: it overestimates the CL capacity of algorithms and relies on unrealistic hyper parameter tuning.
We argue that the evaluation of CL algorithms should focus on assessing the generalizability of their CL capacity to unseen scenarios.
arXiv Detail & Related papers (2024-03-14T03:13:01Z) - Density Distribution-based Learning Framework for Addressing Online
Continual Learning Challenges [4.715630709185073]
We introduce a density distribution-based learning framework for online Continual Learning.
Our framework achieves superior average accuracy and time-space efficiency.
Our method outperforms popular CL approaches by a significant margin.
arXiv Detail & Related papers (2023-11-22T09:21:28Z) - Exploiting CLIP for Zero-shot HOI Detection Requires Knowledge
Distillation at Multiple Levels [52.50670006414656]
We employ CLIP, a large-scale pre-trained vision-language model, for knowledge distillation on multiple levels.
To train our model, CLIP is utilized to generate HOI scores for both global images and local union regions.
The model achieves strong performance, which is even comparable with some fully-supervised and weakly-supervised methods.
arXiv Detail & Related papers (2023-09-10T16:27:54Z) - Real-Time Evaluation in Online Continual Learning: A New Hope [104.53052316526546]
We evaluate current Continual Learning (CL) methods with respect to their computational costs.
A simple baseline outperforms state-of-the-art CL methods under this evaluation.
This surprisingly suggests that the majority of existing CL literature is tailored to a specific class of streams that is not practical.
arXiv Detail & Related papers (2023-02-02T12:21:10Z) - Using Representation Expressiveness and Learnability to Evaluate
Self-Supervised Learning Methods [61.49061000562676]
We introduce Cluster Learnability (CL) to assess learnability.
CL is measured in terms of the performance of a KNN trained to predict labels obtained by clustering the representations with K-means.
We find that CL better correlates with in-distribution model performance than other competing recent evaluation schemes.
arXiv Detail & Related papers (2022-06-02T19:05:13Z) - The CLEAR Benchmark: Continual LEArning on Real-World Imagery [77.98377088698984]
Continual learning (CL) is widely regarded as crucial challenge for lifelong AI.
We introduce CLEAR, the first continual image classification benchmark dataset with a natural temporal evolution of visual concepts.
We find that a simple unsupervised pre-training step can already boost state-of-the-art CL algorithms.
arXiv Detail & Related papers (2022-01-17T09:09:09Z) - Continual Learning for Recurrent Neural Networks: a Review and Empirical
Evaluation [12.27992745065497]
Continual Learning with recurrent neural networks could pave the way to a large number of applications where incoming data is non stationary.
We organize the literature on CL for sequential data processing by providing a categorization of the contributions and a review of the benchmarks.
We propose two new benchmarks for CL with sequential data based on existing datasets, whose characteristics resemble real-world applications.
arXiv Detail & Related papers (2021-03-12T19:25:28Z) - A Survey on Curriculum Learning [48.36129047271622]
Curriculum learning (CL) is a training strategy that trains a machine learning model from easier data to harder data.
As an easy-to-use plug-in, the CL strategy has demonstrated its power in improving the generalization capacity and convergence rate of various models.
arXiv Detail & Related papers (2020-10-25T17:15:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.