Context-Aware Ensemble Learning for Time Series
- URL: http://arxiv.org/abs/2211.16884v1
- Date: Wed, 30 Nov 2022 10:36:13 GMT
- Title: Context-Aware Ensemble Learning for Time Series
- Authors: Arda Fazla, Mustafa Enes Aydin, Orhun Tamyigit, Suleyman Serdar Kozat
- Abstract summary: We introduce a new approach using a meta learner that effectively combines the base model predictions via using a superset of the features that is the union of the base models' feature vectors instead of the predictions themselves.
Our model does not use the predictions of the base models as inputs to a machine learning algorithm, but choose the best possible combination at each time step based on the state of the problem.
- Score: 11.716677452529114
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We investigate ensemble methods for prediction in an online setting. Unlike
all the literature in ensembling, for the first time, we introduce a new
approach using a meta learner that effectively combines the base model
predictions via using a superset of the features that is the union of the base
models' feature vectors instead of the predictions themselves. Here, our model
does not use the predictions of the base models as inputs to a machine learning
algorithm, but choose the best possible combination at each time step based on
the state of the problem. We explore three different constraint spaces for the
ensembling of the base learners that linearly combines the base predictions,
which are convex combinations where the components of the ensembling vector are
all nonnegative and sum up to 1; affine combinations where the weight vector
components are required to sum up to 1; and the unconstrained combinations
where the components are free to take any real value. The constraints are both
theoretically analyzed under known statistics and integrated into the learning
procedure of the meta learner as a part of the optimization in an automated
manner. To show the practical efficiency of the proposed method, we employ a
gradient-boosted decision tree and a multi-layer perceptron separately as the
meta learners. Our framework is generic so that one can use other machine
learning architectures as the ensembler as long as they allow for a custom
differentiable loss for minimization. We demonstrate the learning behavior of
our algorithm on synthetic data and the significant performance improvements
over the conventional methods over various real life datasets, extensively used
in the well-known data competitions. Furthermore, we openly share the source
code of the proposed method to facilitate further research and comparison.
Related papers
- Dynamic Post-Hoc Neural Ensemblers [55.15643209328513]
In this study, we explore employing neural networks as ensemble methods.
Motivated by the risk of learning low-diversity ensembles, we propose regularizing the model by randomly dropping base model predictions.
We demonstrate this approach lower bounds the diversity within the ensemble, reducing overfitting and improving generalization capabilities.
arXiv Detail & Related papers (2024-10-06T15:25:39Z) - Achieving More with Less: A Tensor-Optimization-Powered Ensemble Method [53.170053108447455]
Ensemble learning is a method that leverages weak learners to produce a strong learner.
We design a smooth and convex objective function that leverages the concept of margin, making the strong learner more discriminative.
We then compare our algorithm with random forests of ten times the size and other classical methods across numerous datasets.
arXiv Detail & Related papers (2024-08-06T03:42:38Z) - Proof of Swarm Based Ensemble Learning for Federated Learning
Applications [3.2536767864585663]
In federated learning it is not feasible to apply centralised ensemble learning directly due to privacy concerns.
Most distributed consensus algorithms, such as Byzantine fault tolerance (BFT), do not normally perform well in such applications.
We propose PoSw, a novel distributed consensus algorithm for ensemble learning in a federated setting.
arXiv Detail & Related papers (2022-12-28T13:53:34Z) - Heterogeneous Ensemble Learning for Enhanced Crash Forecasts -- A
Frequentest and Machine Learning based Stacking Framework [0.803552105641624]
In this study, we apply one of the key HEM methods, Stacking, to model crash frequency on five lane undivided segments (5T) of urban and suburban arterials.
The prediction performance of Stacking is compared with parametric statistical models (Poisson and negative binomial) and three state of the art machine learning techniques (Decision tree, random forest, and gradient boosting)
arXiv Detail & Related papers (2022-07-21T19:15:53Z) - HyperImpute: Generalized Iterative Imputation with Automatic Model
Selection [77.86861638371926]
We propose a generalized iterative imputation framework for adaptively and automatically configuring column-wise models.
We provide a concrete implementation with out-of-the-box learners, simulators, and interfaces.
arXiv Detail & Related papers (2022-06-15T19:10:35Z) - Mixed-Integer Optimization with Constraint Learning [4.462264781248437]
We establish a broad methodological foundation for mixed-integer optimization with learned constraints.
We exploit the mixed-integer optimization-representability of many machine learning methods.
We demonstrate the method in both World Food Programme planning and chemotherapy optimization.
arXiv Detail & Related papers (2021-11-04T20:19:55Z) - Optimal Ensemble Construction for Multi-Study Prediction with
Applications to COVID-19 Excess Mortality Estimation [7.02598981483736]
Multi-study ensembling uses a two-stage strategy which fits study-specific models and estimates ensemble weights separately.
This approach ignores the ensemble properties at the model-fitting stage, potentially resulting in a loss of efficiency.
We show that when little data is available for a country before the onset of the pandemic, leveraging data from other countries can substantially improve prediction accuracy.
arXiv Detail & Related papers (2021-09-19T16:52:41Z) - Estimating leverage scores via rank revealing methods and randomization [50.591267188664666]
We study algorithms for estimating the statistical leverage scores of rectangular dense or sparse matrices of arbitrary rank.
Our approach is based on combining rank revealing methods with compositions of dense and sparse randomized dimensionality reduction transforms.
arXiv Detail & Related papers (2021-05-23T19:21:55Z) - Few-Shot Named Entity Recognition: A Comprehensive Study [92.40991050806544]
We investigate three schemes to improve the model generalization ability for few-shot settings.
We perform empirical comparisons on 10 public NER datasets with various proportions of labeled data.
We create new state-of-the-art results on both few-shot and training-free settings.
arXiv Detail & Related papers (2020-12-29T23:43:16Z) - Fast Few-Shot Classification by Few-Iteration Meta-Learning [173.32497326674775]
We introduce a fast optimization-based meta-learning method for few-shot classification.
Our strategy enables important aspects of the base learner objective to be learned during meta-training.
We perform a comprehensive experimental analysis, demonstrating the speed and effectiveness of our approach.
arXiv Detail & Related papers (2020-10-01T15:59:31Z) - StackGenVis: Alignment of Data, Algorithms, and Models for Stacking Ensemble Learning Using Performance Metrics [4.237343083490243]
In machine learning (ML), ensemble methods such as bagging, boosting, and stacking are widely-established approaches.
StackGenVis is a visual analytics system for stacked generalization.
arXiv Detail & Related papers (2020-05-04T15:43:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.