DynED: Dynamic Ensemble Diversification in Data Stream Classification
- URL: http://arxiv.org/abs/2308.10807v2
- Date: Wed, 6 Sep 2023 14:27:17 GMT
- Title: DynED: Dynamic Ensemble Diversification in Data Stream Classification
- Authors: Soheil Abadifard, Sepehr Bakhshi, Sanaz Gheibuni, Fazli Can
- Abstract summary: We present a novel ensemble construction and maintenance approach based on MMR (Maximal Marginal Relevance)
The experimental results on both four real and 11 synthetic datasets demonstrate that the proposed approach provides a higher average mean accuracy compared to the five state-of-the-art baselines.
- Score: 2.990411348977783
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Ensemble methods are commonly used in classification due to their remarkable
performance. Achieving high accuracy in a data stream environment is a
challenging task considering disruptive changes in the data distribution, also
known as concept drift. A greater diversity of ensemble components is known to
enhance prediction accuracy in such settings. Despite the diversity of
components within an ensemble, not all contribute as expected to its overall
performance. This necessitates a method for selecting components that exhibit
high performance and diversity. We present a novel ensemble construction and
maintenance approach based on MMR (Maximal Marginal Relevance) that dynamically
combines the diversity and prediction accuracy of components during the process
of structuring an ensemble. The experimental results on both four real and 11
synthetic datasets demonstrate that the proposed approach (DynED) provides a
higher average mean accuracy compared to the five state-of-the-art baselines.
Related papers
- Understanding the Role of Functional Diversity in Weight-Ensembling with Ingredient Selection and Multidimensional Scaling [7.535219325248997]
We introduce two novel weight-ensembling approaches to study the link between performance dynamics and the nature of how each method decides to apply the functionally diverse components.
We develop a visualization tool to explain how each algorithm explores various domains defined via pairwise-distances to further investigate selection and algorithms' convergence.
arXiv Detail & Related papers (2024-09-04T00:24:57Z) - MDDD: Manifold-based Domain Adaptation with Dynamic Distribution for Non-Deep Transfer Learning in Cross-subject and Cross-session EEG-based Emotion Recognition [11.252832459891566]
We propose a novel non-deep transfer learning method, termed as Manifold-based Domain adaptation with Dynamic Distribution (MDDD)
The experimental results indicate that MDDD outperforms traditional non-deep learning methods, achieving an average improvement of 3.54%.
This suggests that MDDD could be a promising method for enhancing the utility and applicability of aBCIs in real-world scenarios.
arXiv Detail & Related papers (2024-04-24T03:08:25Z) - Imbalanced Data Stream Classification using Dynamic Ensemble Selection [0.0]
This work proposes a novel framework for integrating data pre-processing and dynamic ensemble selection.
The proposed framework was evaluated using six artificially generated data streams with differing imbalance ratios.
According to experimental results, data pre-processing combined with Dynamic Ensemble Selection techniques significantly delivers more accuracy.
arXiv Detail & Related papers (2023-09-17T06:51:29Z) - On the Trade-off of Intra-/Inter-class Diversity for Supervised
Pre-training [72.8087629914444]
We study the impact of the trade-off between the intra-class diversity (the number of samples per class) and the inter-class diversity (the number of classes) of a supervised pre-training dataset.
With the size of the pre-training dataset fixed, the best downstream performance comes with a balance on the intra-/inter-class diversity.
arXiv Detail & Related papers (2023-05-20T16:23:50Z) - CAFE: Learning to Condense Dataset by Aligning Features [72.99394941348757]
We propose a novel scheme to Condense dataset by Aligning FEatures (CAFE)
At the heart of our approach is an effective strategy to align features from the real and synthetic data across various scales.
We validate the proposed CAFE across various datasets, and demonstrate that it generally outperforms the state of the art.
arXiv Detail & Related papers (2022-03-03T05:58:49Z) - Improving robustness and calibration in ensembles with diversity
regularization [1.069533806668766]
We introduce a new diversity regularizer for classification tasks that uses out-of-distribution samples.
We show that regularizing diversity can have a significant impact on calibration and robustness, as well as out-of-distribution detection.
arXiv Detail & Related papers (2022-01-26T12:51:11Z) - Trustworthy Multimodal Regression with Mixture of Normal-inverse Gamma
Distributions [91.63716984911278]
We introduce a novel Mixture of Normal-Inverse Gamma distributions (MoNIG) algorithm, which efficiently estimates uncertainty in principle for adaptive integration of different modalities and produces a trustworthy regression result.
Experimental results on both synthetic and different real-world data demonstrate the effectiveness and trustworthiness of our method on various multimodal regression tasks.
arXiv Detail & Related papers (2021-11-11T14:28:12Z) - CoDA: Contrast-enhanced and Diversity-promoting Data Augmentation for
Natural Language Understanding [67.61357003974153]
We propose a novel data augmentation framework dubbed CoDA.
CoDA synthesizes diverse and informative augmented examples by integrating multiple transformations organically.
A contrastive regularization objective is introduced to capture the global relationship among all the data samples.
arXiv Detail & Related papers (2020-10-16T23:57:03Z) - Neural Ensemble Search for Uncertainty Estimation and Dataset Shift [67.57720300323928]
Ensembles of neural networks achieve superior performance compared to stand-alone networks in terms of accuracy, uncertainty calibration and robustness to dataset shift.
We propose two methods for automatically constructing ensembles with emphvarying architectures.
We show that the resulting ensembles outperform deep ensembles not only in terms of accuracy but also uncertainty calibration and robustness to dataset shift.
arXiv Detail & Related papers (2020-06-15T17:38:15Z) - Diverse Instances-Weighting Ensemble based on Region Drift Disagreement
for Concept Drift Adaptation [40.77597229122878]
We propose a diversity measurement based on whether the ensemble members agree on the probability of a regional distribution change.
An instance-based ensemble learning algorithm, called the diverse instance weighting ensemble (DiwE) is developed to address concept drift for data stream classification problems.
arXiv Detail & Related papers (2020-04-13T07:59:25Z) - Diversity inducing Information Bottleneck in Model Ensembles [73.80615604822435]
In this paper, we target the problem of generating effective ensembles of neural networks by encouraging diversity in prediction.
We explicitly optimize a diversity inducing adversarial loss for learning latent variables and thereby obtain diversity in the output predictions necessary for modeling multi-modal data.
Compared to the most competitive baselines, we show significant improvements in classification accuracy, under a shift in the data distribution.
arXiv Detail & Related papers (2020-03-10T03:10:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.