A Broad Ensemble Learning System for Drifting Stream Classification
- URL: http://arxiv.org/abs/2110.03540v1
- Date: Thu, 7 Oct 2021 15:01:33 GMT
- Title: A Broad Ensemble Learning System for Drifting Stream Classification
- Authors: Sepehr Bakhshi, Pouya Ghahramanian, Hamed Bonab, and Fazli Can
- Abstract summary: We propose a Broad Ensemble Learning System (BELS) for stream classification with concept drift.
BELS uses a novel updating method that greatly improves best-in-class model accuracy.
We show that our proposed method improves on average 44% compared to BLS, and 29% compared to other competitive baselines.
- Score: 3.087840197124265
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Data stream classification has become a major research topic due to the
increase in temporal data. One of the biggest hurdles of data stream
classification is the development of algorithms that deal with evolving data,
also known as concept drifts. As data changes over time, static prediction
models lose their validity. Adapting to concept drifts provides more robust and
better performing models. The Broad Learning System (BLS) is an effective broad
neural architecture recently developed for incremental learning. BLS cannot
provide instant response since it requires huge data chunks and is unable to
handle concept drifts. We propose a Broad Ensemble Learning System (BELS) for
stream classification with concept drift. BELS uses a novel updating method
that greatly improves best-in-class model accuracy. It employs a dynamic output
ensemble layer to address the limitations of BLS. We present its mathematical
derivation, provide comprehensive experiments with 11 datasets that demonstrate
the adaptability of our model, including a comparison of our model with BLS,
and provide parameter and robustness analysis on several drifting streams,
showing that it statistically significantly outperforms seven state-of-the-art
baselines. We show that our proposed method improves on average 44% compared to
BLS, and 29% compared to other competitive baselines.
Related papers
- Online-BLS: An Accurate and Efficient Online Broad Learning System for Data Stream Classification [52.251569042852815]
We introduce an online broad learning system framework with closed-form solutions for each online update.
We design an effective weight estimation algorithm and an efficient online updating strategy.
Our framework is naturally extended to data stream scenarios with concept drift and exceeds state-of-the-art baselines.
arXiv Detail & Related papers (2025-01-28T13:21:59Z) - A New Perspective on Time Series Anomaly Detection: Faster Patch-based Broad Learning System [59.38402187365612]
Time series anomaly detection (TSAD) has been a research hotspot in both academia and industry in recent years.
Deep learning is not required for TSAD due to limitations such as slow deep learning speed.
We propose Contrastive Patch-based Broad Learning System (CBLS)
arXiv Detail & Related papers (2024-12-07T01:58:18Z) - Act Now: A Novel Online Forecasting Framework for Large-Scale Streaming Data [17.121851529311368]
Existing online forecasting methods have the following issues.
They do not consider the update frequency of streaming data.
Eliminating information leakage can exacerbate concept drift.
Existing GPU devices cannot support online learning of large-scale streaming data.
arXiv Detail & Related papers (2024-11-28T01:39:45Z) - Drift-Resilient TabPFN: In-Context Learning Temporal Distribution Shifts on Tabular Data [39.40116554523575]
We present Drift-Resilient TabPFN, a fresh approach based on In-Context Learning with a Prior-Data Fitted Network.
It learns to approximate Bayesian inference on synthetic datasets drawn from a prior.
It improves accuracy from 0.688 to 0.744 and ROC AUC from 0.786 to 0.832 while maintaining stronger calibration.
arXiv Detail & Related papers (2024-11-15T23:49:23Z) - A Bayesian Approach to Data Point Selection [24.98069363998565]
Data point selection (DPS) is becoming a critical topic in deep learning.
Existing approaches to DPS are predominantly based on a bi-level optimisation (BLO) formulation.
We propose a novel Bayesian approach to DPS.
arXiv Detail & Related papers (2024-11-06T09:04:13Z) - Fisher Information-based Efficient Curriculum Federated Learning with Large Language Models [43.26028399395612]
We propose a Fisher Information-based Efficient Curriculum Federated Learning framework (FibecFed) with two novel methods.
First, we propose a fisher information-based method to adaptively sample data within each device to improve the effectiveness of the FL fine-tuning process.
Second, we dynamically select the proper layers for global aggregation and sparse parameters for local update with LoRA.
arXiv Detail & Related papers (2024-09-30T18:12:18Z) - Robust Learning with Progressive Data Expansion Against Spurious
Correlation [65.83104529677234]
We study the learning process of a two-layer nonlinear convolutional neural network in the presence of spurious features.
Our analysis suggests that imbalanced data groups and easily learnable spurious features can lead to the dominance of spurious features during the learning process.
We propose a new training algorithm called PDE that efficiently enhances the model's robustness for a better worst-group performance.
arXiv Detail & Related papers (2023-06-08T05:44:06Z) - Bilevel Online Deep Learning in Non-stationary Environment [4.565872584112864]
Bilevel Online Deep Learning (BODL) framework combines bilevel optimization strategy and online ensemble classifier.
When the concept drift is detected, our BODL algorithm can adaptively update the model parameters via bilevel optimization and then circumvent the large drift and encourage positive transfer.
arXiv Detail & Related papers (2022-01-25T11:05:51Z) - Hyperparameter-free Continuous Learning for Domain Classification in
Natural Language Understanding [60.226644697970116]
Domain classification is the fundamental task in natural language understanding (NLU)
Most existing continual learning approaches suffer from low accuracy and performance fluctuation.
We propose a hyper parameter-free continual learning model for text data that can stably produce high performance under various environments.
arXiv Detail & Related papers (2022-01-05T02:46:16Z) - RIFLE: Backpropagation in Depth for Deep Transfer Learning through
Re-Initializing the Fully-connected LayEr [60.07531696857743]
Fine-tuning the deep convolution neural network(CNN) using a pre-trained model helps transfer knowledge learned from larger datasets to the target task.
We propose RIFLE - a strategy that deepens backpropagation in transfer learning settings.
RIFLE brings meaningful updates to the weights of deep CNN layers and improves low-level feature learning.
arXiv Detail & Related papers (2020-07-07T11:27:43Z) - Learnable Bernoulli Dropout for Bayesian Deep Learning [53.79615543862426]
Learnable Bernoulli dropout (LBD) is a new model-agnostic dropout scheme that considers the dropout rates as parameters jointly optimized with other model parameters.
LBD leads to improved accuracy and uncertainty estimates in image classification and semantic segmentation.
arXiv Detail & Related papers (2020-02-12T18:57:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.