Online Learning With Adaptive Rebalancing in Nonstationary Environments
- URL: http://arxiv.org/abs/2009.11942v1
- Date: Thu, 24 Sep 2020 20:40:04 GMT
- Title: Online Learning With Adaptive Rebalancing in Nonstationary Environments
- Authors: Kleanthis Malialis and Christos G. Panayiotou and Marios M. Polycarpou
- Abstract summary: We provide new insights into learning from nonstationary and imbalanced data in online learning.
We propose the novel Adaptive REBAlancing (AREBA) algorithm that selectively includes in the training set a subset of the majority and minority examples.
- Score: 11.501721946030779
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: An enormous and ever-growing volume of data is nowadays becoming available in
a sequential fashion in various real-world applications. Learning in
nonstationary environments constitutes a major challenge, and this problem
becomes orders of magnitude more complex in the presence of class imbalance. We
provide new insights into learning from nonstationary and imbalanced data in
online learning, a largely unexplored area. We propose the novel Adaptive
REBAlancing (AREBA) algorithm that selectively includes in the training set a
subset of the majority and minority examples that appeared so far, while at its
heart lies an adaptive mechanism to continually maintain the class balance
between the selected examples. We compare AREBA with strong baselines and other
state-of-the-art algorithms and perform extensive experimental work in
scenarios with various class imbalance rates and different concept drift types
on both synthetic and real-world data. AREBA significantly outperforms the rest
with respect to both learning speed and learning quality. Our code is made
publicly available to the scientific community.
Related papers
- Efficient Imitation Learning with Conservative World Models [54.52140201148341]
We tackle the problem of policy learning from expert demonstrations without a reward function.
We re-frame imitation learning as a fine-tuning problem, rather than a pure reinforcement learning one.
arXiv Detail & Related papers (2024-05-21T20:53:18Z) - A Survey of Deep Long-Tail Classification Advancements [1.6233132273470656]
Many data distributions in the real world are hardly uniform. Instead, skewed and long-tailed distributions of various kinds are commonly observed.
This poses an interesting problem for machine learning, where most algorithms assume or work well with uniformly distributed data.
The problem is further exacerbated by current state-of-the-art deep learning models requiring large volumes of training data.
arXiv Detail & Related papers (2024-04-24T01:59:02Z) - Time Will Change Things: An Empirical Study on Dynamic Language
Understanding in Social Media Classification [5.075802830306718]
We empirically study social media NLU in a dynamic setup, where models are trained on the past data and test on the future.
We show that auto-encoding and pseudo-labeling collaboratively show the best robustness in dynamicity.
arXiv Detail & Related papers (2022-10-06T12:18:28Z) - Nonstationary data stream classification with online active learning and
siamese neural networks [11.501721946030779]
An emerging need for online learning methods that train predictive models on-the-fly.
A series of open challenges, however, hinder their deployment in practice.
We propose the ActiSiamese algorithm, which addresses these challenges by combining online active learning, siamese networks, and a multi-queue memory.
arXiv Detail & Related papers (2022-10-03T17:16:03Z) - Deep Reinforcement Learning for Multi-class Imbalanced Training [64.9100301614621]
We introduce an imbalanced classification framework, based on reinforcement learning, for training extremely imbalanced data sets.
We formulate a custom reward function and episode-training procedure, specifically with the added capability of handling multi-class imbalanced training.
Using real-world clinical case studies, we demonstrate that our proposed framework outperforms current state-of-the-art imbalanced learning methods.
arXiv Detail & Related papers (2022-05-24T13:39:59Z) - ASE: Anomaly Scoring Based Ensemble Learning for Imbalanced Datasets [3.214208422566496]
We come up with a bagging ensemble learning framework based on an anomaly detection scoring system.
We test out that our ensemble learning model can dramatically improve performance of base estimators.
arXiv Detail & Related papers (2022-03-21T07:20:41Z) - Efficient Model-based Multi-agent Reinforcement Learning via Optimistic
Equilibrium Computation [93.52573037053449]
H-MARL (Hallucinated Multi-Agent Reinforcement Learning) learns successful equilibrium policies after a few interactions with the environment.
We demonstrate our approach experimentally on an autonomous driving simulation benchmark.
arXiv Detail & Related papers (2022-03-14T17:24:03Z) - CMW-Net: Learning a Class-Aware Sample Weighting Mapping for Robust Deep
Learning [55.733193075728096]
Modern deep neural networks can easily overfit to biased training data containing corrupted labels or class imbalance.
Sample re-weighting methods are popularly used to alleviate this data bias issue.
We propose a meta-model capable of adaptively learning an explicit weighting scheme directly from data.
arXiv Detail & Related papers (2022-02-11T13:49:51Z) - Quasi-Global Momentum: Accelerating Decentralized Deep Learning on
Heterogeneous Data [77.88594632644347]
Decentralized training of deep learning models is a key element for enabling data privacy and on-device learning over networks.
In realistic learning scenarios, the presence of heterogeneity across different clients' local datasets poses an optimization challenge.
We propose a novel momentum-based method to mitigate this decentralized training difficulty.
arXiv Detail & Related papers (2021-02-09T11:27:14Z) - Data-efficient Online Classification with Siamese Networks and Active
Learning [11.501721946030779]
We investigate learning from limited labelled, nonstationary and imbalanced data in online classification.
We propose a learning method that synergistically combines siamese neural networks and active learning.
Our study shows that the proposed method is robust to data nonstationarity and imbalance, and significantly outperforms baselines and state-of-the-art algorithms in terms of both learning speed and performance.
arXiv Detail & Related papers (2020-10-04T19:07:19Z) - Long-Tailed Recognition Using Class-Balanced Experts [128.73438243408393]
We propose an ensemble of class-balanced experts that combines the strength of diverse classifiers.
Our ensemble of class-balanced experts reaches results close to state-of-the-art and an extended ensemble establishes a new state-of-the-art on two benchmarks for long-tailed recognition.
arXiv Detail & Related papers (2020-04-07T20:57:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.