Which is the best model for my data?
- URL: http://arxiv.org/abs/2210.14687v1
- Date: Wed, 26 Oct 2022 13:15:43 GMT
- Title: Which is the best model for my data?
- Authors: Gonzalo N\'apoles and Isel Grau and \c{C}i\c{c}ek G\"uven and
Or\c{c}un \"Ozdemir and Yamisleydi Salgueiro
- Abstract summary: The proposed meta-learning approach relies on machine learning and involves four major steps.
We present a collection of 62 meta-features that address the problem of information cancellation when aggregation measure values involving positive and negative measurements.
We show that our meta-learning approach can correctly predict an optimal model for 91% of the synthetic datasets and for 87% of the real-world datasets.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In this paper, we tackle the problem of selecting the optimal model for a
given structured pattern classification dataset. In this context, a model can
be understood as a classifier and a hyperparameter configuration. The proposed
meta-learning approach purely relies on machine learning and involves four
major steps. Firstly, we present a concise collection of 62 meta-features that
address the problem of information cancellation when aggregation measure values
involving positive and negative measurements. Secondly, we describe two
different approaches for synthetic data generation intending to enlarge the
training data. Thirdly, we fit a set of pre-defined classification models for
each classification problem while optimizing their hyperparameters using grid
search. The goal is to create a meta-dataset such that each row denotes a
multilabel instance describing a specific problem. The features of these
meta-instances denote the statistical properties of the generated datasets,
while the labels encode the grid search results as binary vectors such that
best-performing models are positively labeled. Finally, we tackle the model
selection problem with several multilabel classifiers, including a
Convolutional Neural Network designed to handle tabular data. The simulation
results show that our meta-learning approach can correctly predict an optimal
model for 91% of the synthetic datasets and for 87% of the real-world datasets.
Furthermore, we noticed that most meta-classifiers produced better results when
using our meta-features. Overall, our proposal differs from other meta-learning
approaches since it tackles the algorithm selection and hyperparameter tuning
problems in a single step. Toward the end, we perform a feature importance
analysis to determine which statistical features drive the model selection
mechanism.
Related papers
- An incremental preference elicitation-based approach to learning potentially non-monotonic preferences in multi-criteria sorting [53.36437745983783]
We first construct a max-margin optimization-based model to model potentially non-monotonic preferences.
We devise information amount measurement methods and question selection strategies to pinpoint the most informative alternative in each iteration.
Two incremental preference elicitation-based algorithms are developed to learn potentially non-monotonic preferences.
arXiv Detail & Related papers (2024-09-04T14:36:20Z) - DsDm: Model-Aware Dataset Selection with Datamodels [81.01744199870043]
Standard practice is to filter for examples that match human notions of data quality.
We find that selecting according to similarity with "high quality" data sources may not increase (and can even hurt) performance compared to randomly selecting data.
Our framework avoids handpicked notions of data quality, and instead models explicitly how the learning process uses train datapoints to predict on the target tasks.
arXiv Detail & Related papers (2024-01-23T17:22:00Z) - Quick-Tune: Quickly Learning Which Pretrained Model to Finetune and How [62.467716468917224]
We propose a methodology that jointly searches for the optimal pretrained model and the hyperparameters for finetuning it.
Our method transfers knowledge about the performance of many pretrained models on a series of datasets.
We empirically demonstrate that our resulting approach can quickly select an accurate pretrained model for a new dataset.
arXiv Detail & Related papers (2023-06-06T16:15:26Z) - Automatic learning algorithm selection for classification via
convolutional neural networks [0.0]
The goal of this study is to learn the inherent structure of the data without identifying meta-features.
Experiments with simulated datasets show that the proposed approach achieves nearly perfect performance in identifying linear and nonlinear patterns.
arXiv Detail & Related papers (2023-05-16T01:57:01Z) - Improving Group Lasso for high-dimensional categorical data [0.90238471756546]
Group Lasso is a well known efficient algorithm for selection continuous or categorical variables.
We propose a two-step procedure to obtain a sparse solution of the Group Lasso.
We show that our method performs better than the state of the art algorithms with respect to the prediction accuracy or model dimension.
arXiv Detail & Related papers (2022-10-25T13:43:57Z) - Low-rank Dictionary Learning for Unsupervised Feature Selection [11.634317251468968]
We introduce a novel unsupervised feature selection approach by applying dictionary learning ideas in a low-rank representation.
A unified objective function for unsupervised feature selection is proposed in a sparse way by an $ell_2,1$-norm regularization.
Our experimental findings reveal that the proposed method outperforms the state-of-the-art algorithm.
arXiv Detail & Related papers (2021-06-21T13:39:10Z) - Auto-weighted Multi-view Feature Selection with Graph Optimization [90.26124046530319]
We propose a novel unsupervised multi-view feature selection model based on graph learning.
The contributions are threefold: (1) during the feature selection procedure, the consensus similarity graph shared by different views is learned.
Experiments on various datasets demonstrate the superiority of the proposed method compared with the state-of-the-art methods.
arXiv Detail & Related papers (2021-04-11T03:25:25Z) - Adaptive Graph-based Generalized Regression Model for Unsupervised
Feature Selection [11.214334712819396]
How to select the uncorrelated and discriminative features is the key problem of unsupervised feature selection.
We present a novel generalized regression model imposed by an uncorrelated constraint and the $ell_2,1$-norm regularization.
It can simultaneously select the uncorrelated and discriminative features as well as reduce the variance of these data points belonging to the same neighborhood.
arXiv Detail & Related papers (2020-12-27T09:07:26Z) - Machine learning with incomplete datasets using multi-objective
optimization models [1.933681537640272]
We propose an online approach to handle missing values while a classification model is learnt.
We develop a multi-objective optimization model with two objective functions for imputation and model selection.
We use an evolutionary algorithm based on NSGA II to find the optimal solutions.
arXiv Detail & Related papers (2020-12-04T03:44:33Z) - Semi-Supervised Learning with Meta-Gradient [123.26748223837802]
We propose a simple yet effective meta-learning algorithm in semi-supervised learning.
We find that the proposed algorithm performs favorably against state-of-the-art methods.
arXiv Detail & Related papers (2020-07-08T08:48:56Z) - Pre-Trained Models for Heterogeneous Information Networks [57.78194356302626]
We propose a self-supervised pre-training and fine-tuning framework, PF-HIN, to capture the features of a heterogeneous information network.
PF-HIN consistently and significantly outperforms state-of-the-art alternatives on each of these tasks, on four datasets.
arXiv Detail & Related papers (2020-07-07T03:36:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.