Continual Learning for Human State Monitoring
- URL: http://arxiv.org/abs/2207.00010v1
- Date: Wed, 29 Jun 2022 19:23:13 GMT
- Title: Continual Learning for Human State Monitoring
- Authors: Federico Matteoni, Andrea Cossu, Claudio Gallicchio, Vincenzo
Lomonaco, Davide Bacciu
- Abstract summary: We propose two new CL benchmarks for Human State Monitoring.
We carefully designed the benchmarks to mirror real-world environments in which new subjects are continuously added.
We conducted an empirical evaluation to assess the ability of popular CL strategies to mitigate forgetting in our benchmarks.
- Score: 20.8311956676327
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Continual Learning (CL) on time series data represents a promising but
under-studied avenue for real-world applications. We propose two new CL
benchmarks for Human State Monitoring. We carefully designed the benchmarks to
mirror real-world environments in which new subjects are continuously added. We
conducted an empirical evaluation to assess the ability of popular CL
strategies to mitigate forgetting in our benchmarks. Our results show that,
possibly due to the domain-incremental properties of our benchmarks, forgetting
can be easily tackled even with a simple finetuning and that existing
strategies struggle in accumulating knowledge over a fixed, held-out, test
subject.
Related papers
- Position: LLM Unlearning Benchmarks are Weak Measures of Progress [31.957968729934745]
We find that existing benchmarks provide an overly optimistic and potentially misleading view on the effectiveness of candidate unlearning methods.
We identify that existing benchmarks are particularly vulnerable to modifications that introduce even loose dependencies between the forget and retain information.
arXiv Detail & Related papers (2024-10-03T18:07:25Z) - Automating Dataset Updates Towards Reliable and Timely Evaluation of Large Language Models [81.27391252152199]
Large language models (LLMs) have achieved impressive performance across various natural language benchmarks.
We propose to automate dataset updating and provide systematic analysis regarding its effectiveness.
There are two updating strategies: 1) mimicking strategy to generate similar samples based on original data, and 2) extending strategy that further expands existing samples.
arXiv Detail & Related papers (2024-02-19T07:15:59Z) - Unsupervised Temporal Action Localization via Self-paced Incremental
Learning [57.55765505856969]
We present a novel self-paced incremental learning model to enhance clustering and localization training simultaneously.
We design two (constant- and variable- speed) incremental instance learning strategies for easy-to-hard model training, thus ensuring the reliability of these video pseudolabels.
arXiv Detail & Related papers (2023-12-12T16:00:55Z) - Don't Make Your LLM an Evaluation Benchmark Cheater [142.24553056600627]
Large language models(LLMs) have greatly advanced the frontiers of artificial intelligence, attaining remarkable improvement in model capacity.
To assess the model performance, a typical approach is to construct evaluation benchmarks for measuring the ability level of LLMs.
We discuss the potential risk and impact of inappropriately using evaluation benchmarks and misleadingly interpreting the evaluation results.
arXiv Detail & Related papers (2023-11-03T14:59:54Z) - Re-Benchmarking Pool-Based Active Learning for Binary Classification [27.034593234956713]
Active learning is a paradigm that significantly enhances the performance of machine learning models when acquiring labeled data.
While several benchmarks exist for evaluating active learning strategies, their findings exhibit some misalignment.
This discrepancy motivates us to develop a transparent and reproducible benchmark for the community.
arXiv Detail & Related papers (2023-06-15T08:47:50Z) - A Large-scale Study of Spatiotemporal Representation Learning with a New
Benchmark on Action Recognition [14.226201098201244]
BEAR is a collection of 18 video datasets grouped into 5 categories (anomaly, gesture, daily, sports, and instructional)
We thoroughly evaluate 6 commontemporal models pre-trained by both supervised and self-supervised learning.
Our observation suggests that current state-of-the-art cannot solidly guarantee high performance on datasets close to real-world applications.
arXiv Detail & Related papers (2023-03-23T17:58:05Z) - From MNIST to ImageNet and Back: Benchmarking Continual Curriculum
Learning [9.104068727716294]
Continual learning (CL) is one of the most promising trends in machine learning research.
We introduce two novel CL benchmarks that involve multiple heterogeneous tasks from six image datasets.
We additionally structure our benchmarks so that tasks are presented in increasing and decreasing order of complexity.
arXiv Detail & Related papers (2023-03-16T18:11:19Z) - Real-Time Evaluation in Online Continual Learning: A New Hope [104.53052316526546]
We evaluate current Continual Learning (CL) methods with respect to their computational costs.
A simple baseline outperforms state-of-the-art CL methods under this evaluation.
This surprisingly suggests that the majority of existing CL literature is tailored to a specific class of streams that is not practical.
arXiv Detail & Related papers (2023-02-02T12:21:10Z) - CLAD: A realistic Continual Learning benchmark for Autonomous Driving [33.95470797472666]
This paper describes the design and the ideas motivating a new Continual Learning benchmark for Autonomous Driving.
The benchmark uses SODA10M, a recently released large-scale dataset that concerns autonomous driving related problems.
We introduce CLAD-C, an online classification benchmark realised through a chronological data stream that poses both class and domain incremental challenges.
We examine the inherent difficulties and challenges posed by the benchmark, through a survey of the techniques and methods used by the top-3 participants in a CLAD-challenge workshop at ICCV 2021.
arXiv Detail & Related papers (2022-10-07T12:08:25Z) - Using Representation Expressiveness and Learnability to Evaluate
Self-Supervised Learning Methods [61.49061000562676]
We introduce Cluster Learnability (CL) to assess learnability.
CL is measured in terms of the performance of a KNN trained to predict labels obtained by clustering the representations with K-means.
We find that CL better correlates with in-distribution model performance than other competing recent evaluation schemes.
arXiv Detail & Related papers (2022-06-02T19:05:13Z) - The CLEAR Benchmark: Continual LEArning on Real-World Imagery [77.98377088698984]
Continual learning (CL) is widely regarded as crucial challenge for lifelong AI.
We introduce CLEAR, the first continual image classification benchmark dataset with a natural temporal evolution of visual concepts.
We find that a simple unsupervised pre-training step can already boost state-of-the-art CL algorithms.
arXiv Detail & Related papers (2022-01-17T09:09:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.