Instance exploitation for learning temporary concepts from sparsely
labeled drifting data streams
- URL: http://arxiv.org/abs/2009.09382v1
- Date: Sun, 20 Sep 2020 08:11:43 GMT
- Title: Instance exploitation for learning temporary concepts from sparsely
labeled drifting data streams
- Authors: {\L}ukasz Korycki and Bartosz Krawczyk
- Abstract summary: Continual learning from streaming data sources becomes more and more popular.
dealing with dynamic and everlasting problems poses new challenges.
One of the most crucial limitations is that we cannot assume having access to a finite and complete data set.
- Score: 15.49323098362628
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Continual learning from streaming data sources becomes more and more popular
due to the increasing number of online tools and systems. Dealing with dynamic
and everlasting problems poses new challenges for which traditional batch-based
offline algorithms turn out to be insufficient in terms of computational time
and predictive performance. One of the most crucial limitations is that we
cannot assume having access to a finite and complete data set - we always have
to be ready for new data that may complement our model. This poses a critical
problem of providing labels for potentially unbounded streams. In the real
world, we are forced to deal with very strict budget limitations, therefore, we
will most likely face the scarcity of annotated instances, which are essential
in supervised learning. In our work, we emphasize this problem and propose a
novel instance exploitation technique. We show that when: (i) data is
characterized by temporary non-stationary concepts, and (ii) there are very few
labels spanned across a long time horizon, it is actually better to risk
overfitting and adapt models more aggressively by exploiting the only labeled
instances we have, instead of sticking to a standard learning mode and
suffering from severe underfitting. We present different strategies and
configurations for our methods, as well as an ensemble algorithm that attempts
to maintain a sweet spot between risky and normal adaptation. Finally, we
conduct a complex in-depth comparative analysis of our methods, using
state-of-the-art streaming algorithms relevant to the given problem.
Related papers
- Federated Continual Learning Goes Online: Uncertainty-Aware Memory Management for Vision Tasks and Beyond [13.867793835583463]
We propose an uncertainty-aware memory-based approach to solve catastrophic forgetting.
We retrieve samples with specific characteristics, and - by retraining the model on such samples - we demonstrate the potential of this approach.
arXiv Detail & Related papers (2024-05-29T09:29:39Z) - Towards Robust Continual Learning with Bayesian Adaptive Moment Regularization [51.34904967046097]
Continual learning seeks to overcome the challenge of catastrophic forgetting, where a model forgets previously learnt information.
We introduce a novel prior-based method that better constrains parameter growth, reducing catastrophic forgetting.
Results show that BAdam achieves state-of-the-art performance for prior-based methods on challenging single-headed class-incremental experiments.
arXiv Detail & Related papers (2023-09-15T17:10:51Z) - Large-scale Pre-trained Models are Surprisingly Strong in Incremental Novel Class Discovery [76.63807209414789]
We challenge the status quo in class-iNCD and propose a learning paradigm where class discovery occurs continuously and truly unsupervisedly.
We propose simple baselines, composed of a frozen PTM backbone and a learnable linear classifier, that are not only simple to implement but also resilient under longer learning scenarios.
arXiv Detail & Related papers (2023-03-28T13:47:16Z) - Learning from Data Streams: An Overview and Update [1.5076964620370268]
We reformulate the fundamental definitions and settings of supervised data-stream learning.
We take a fresh look at what constitutes a supervised data-stream learning task.
Our main emphasis is that learning from data streams does not impose a single-pass or online-learning approach.
arXiv Detail & Related papers (2022-12-30T14:01:41Z) - Nonstationary data stream classification with online active learning and
siamese neural networks [11.501721946030779]
An emerging need for online learning methods that train predictive models on-the-fly.
A series of open challenges, however, hinder their deployment in practice.
We propose the ActiSiamese algorithm, which addresses these challenges by combining online active learning, siamese networks, and a multi-queue memory.
arXiv Detail & Related papers (2022-10-03T17:16:03Z) - Augmented Bilinear Network for Incremental Multi-Stock Time-Series
Classification [83.23129279407271]
We propose a method to efficiently retain the knowledge available in a neural network pre-trained on a set of securities.
In our method, the prior knowledge encoded in a pre-trained neural network is maintained by keeping existing connections fixed.
This knowledge is adjusted for the new securities by a set of augmented connections, which are optimized using the new data.
arXiv Detail & Related papers (2022-07-23T18:54:10Z) - Continual Few-shot Relation Learning via Embedding Space Regularization
and Data Augmentation [4.111899441919165]
It is necessary for the model to learn novel relational patterns with very few labeled data while avoiding catastrophic forgetting of previous task knowledge.
We propose a novel method based on embedding space regularization and data augmentation.
Our method generalizes to new few-shot tasks and avoids catastrophic forgetting of previous tasks by enforcing extra constraints on the relational embeddings and by adding extra relevant data in a self-supervised manner.
arXiv Detail & Related papers (2022-03-04T05:19:09Z) - Mining Drifting Data Streams on a Budget: Combining Active Learning with
Self-Labeling [6.436899373275926]
We propose a novel framework for mining drifting data streams on a budget, by combining information coming from active learning and self-labeling.
We introduce several strategies that can take advantage of both intelligent instance selection and semi-supervised procedures, while taking into account the potential presence of concept drift.
arXiv Detail & Related papers (2021-12-21T07:19:35Z) - Constrained Learning with Non-Convex Losses [119.8736858597118]
Though learning has become a core technology of modern information processing, there is now ample evidence that it can lead to biased, unsafe, and prejudiced solutions.
arXiv Detail & Related papers (2021-03-08T23:10:33Z) - Semi-Supervised Learning with Meta-Gradient [123.26748223837802]
We propose a simple yet effective meta-learning algorithm in semi-supervised learning.
We find that the proposed algorithm performs favorably against state-of-the-art methods.
arXiv Detail & Related papers (2020-07-08T08:48:56Z) - Differentiable Causal Discovery from Interventional Data [141.41931444927184]
We propose a theoretically-grounded method based on neural networks that can leverage interventional data.
We show that our approach compares favorably to the state of the art in a variety of settings.
arXiv Detail & Related papers (2020-07-03T15:19:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.