Conceptually Diverse Base Model Selection for Meta-Learners in Concept
Drifting Data Streams
- URL: http://arxiv.org/abs/2111.14520v1
- Date: Mon, 29 Nov 2021 13:18:53 GMT
- Title: Conceptually Diverse Base Model Selection for Meta-Learners in Concept
Drifting Data Streams
- Authors: Helen McKay, Nathan Griffiths, Phillip Taylor
- Abstract summary: We present a novel approach for estimating the conceptual similarity of base models, which is calculated using the Principal Angles (PAs) between their underlying subspaces.
We evaluate these methods against thresholding using common ensemble pruning metrics, namely predictive performance and Mutual Information (MI) in the context of online Transfer Learning (TL)
Our results show that conceptual similarity thresholding has a reduced computational overhead, and yet yields comparable predictive performance to thresholding using predictive performance and MI.
- Score: 3.0938904602244355
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Meta-learners and ensembles aim to combine a set of relevant yet diverse base
models to improve predictive performance. However, determining an appropriate
set of base models is challenging, especially in online environments where the
underlying distribution of data can change over time. In this paper, we present
a novel approach for estimating the conceptual similarity of base models, which
is calculated using the Principal Angles (PAs) between their underlying
subspaces. We propose two methods that use conceptual similarity as a metric to
obtain a relevant yet diverse subset of base models: (i) parameterised
threshold culling and (ii) parameterless conceptual clustering. We evaluate
these methods against thresholding using common ensemble pruning metrics,
namely predictive performance and Mutual Information (MI), in the context of
online Transfer Learning (TL), using both synthetic and real-world data. Our
results show that conceptual similarity thresholding has a reduced
computational overhead, and yet yields comparable predictive performance to
thresholding using predictive performance and MI. Furthermore, conceptual
clustering achieves similar predictive performances without requiring
parameterisation, and achieves this with lower computational overhead than
thresholding using predictive performance and MI when the number of base models
becomes large.
Related papers
- Beyond explaining: XAI-based Adaptive Learning with SHAP Clustering for
Energy Consumption Prediction [0.0]
We introduce a three-stage process: obtaining SHAP values to explain model predictions, clustering SHAP values to identify distinct patterns and outliers, and refining the model based on the derived SHAP clustering characteristics.
Our experiments demonstrate the effectiveness of our approach in both task types, resulting in improved predictive performance and interpretable model explanations.
arXiv Detail & Related papers (2024-02-07T15:58:51Z) - Online Variational Sequential Monte Carlo [49.97673761305336]
We build upon the variational sequential Monte Carlo (VSMC) method, which provides computationally efficient and accurate model parameter estimation and Bayesian latent-state inference.
Online VSMC is capable of performing efficiently, entirely on-the-fly, both parameter estimation and particle proposal adaptation.
arXiv Detail & Related papers (2023-12-19T21:45:38Z) - Aggregation Weighting of Federated Learning via Generalization Bound
Estimation [65.8630966842025]
Federated Learning (FL) typically aggregates client model parameters using a weighting approach determined by sample proportions.
We replace the aforementioned weighting method with a new strategy that considers the generalization bounds of each local model.
arXiv Detail & Related papers (2023-11-10T08:50:28Z) - Consensus-Adaptive RANSAC [104.87576373187426]
We propose a new RANSAC framework that learns to explore the parameter space by considering the residuals seen so far via a novel attention layer.
The attention mechanism operates on a batch of point-to-model residuals, and updates a per-point estimation state to take into account the consensus found through a lightweight one-step transformer.
arXiv Detail & Related papers (2023-07-26T08:25:46Z) - Multi-Task Learning Regression via Convex Clustering [0.0]
We propose an MTL method with a centroid parameter representing a cluster center of the task.
We show the effectiveness of the proposed method through Monte Carlo simulations and applications to real data.
arXiv Detail & Related papers (2023-04-26T07:25:21Z) - Evaluating Representations with Readout Model Switching [18.475866691786695]
In this paper, we propose to use the Minimum Description Length (MDL) principle to devise an evaluation metric.
We design a hybrid discrete and continuous-valued model space for the readout models and employ a switching strategy to combine their predictions.
The proposed metric can be efficiently computed with an online method and we present results for pre-trained vision encoders of various architectures.
arXiv Detail & Related papers (2023-02-19T14:08:01Z) - Federated Learning Aggregation: New Robust Algorithms with Guarantees [63.96013144017572]
Federated learning has been recently proposed for distributed model training at the edge.
This paper presents a complete general mathematical convergence analysis to evaluate aggregation strategies in a federated learning framework.
We derive novel aggregation algorithms which are able to modify their model architecture by differentiating client contributions according to the value of their losses.
arXiv Detail & Related papers (2022-05-22T16:37:53Z) - Distributional Depth-Based Estimation of Object Articulation Models [21.046351215949525]
We propose a method that efficiently learns distributions over articulation model parameters directly from depth images.
Our core contributions include a novel representation for distributions over rigid body transformations.
We introduce a novel deep learning based approach, DUST-net, that performs category-independent articulation model estimation.
arXiv Detail & Related papers (2021-08-12T17:44:51Z) - MINIMALIST: Mutual INformatIon Maximization for Amortized Likelihood
Inference from Sampled Trajectories [61.3299263929289]
Simulation-based inference enables learning the parameters of a model even when its likelihood cannot be computed in practice.
One class of methods uses data simulated with different parameters to infer an amortized estimator for the likelihood-to-evidence ratio.
We show that this approach can be formulated in terms of mutual information between model parameters and simulated data.
arXiv Detail & Related papers (2021-06-03T12:59:16Z) - Control as Hybrid Inference [62.997667081978825]
We present an implementation of CHI which naturally mediates the balance between iterative and amortised inference.
We verify the scalability of our algorithm on a continuous control benchmark, demonstrating that it outperforms strong model-free and model-based baselines.
arXiv Detail & Related papers (2020-07-11T19:44:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.