Explainable Model-specific Algorithm Selection for Multi-Label
Classification
- URL: http://arxiv.org/abs/2211.11227v1
- Date: Mon, 21 Nov 2022 07:42:11 GMT
- Title: Explainable Model-specific Algorithm Selection for Multi-Label
Classification
- Authors: Ana Kostovska, Carola Doerr, Sa\v{s}o D\v{z}eroski, Dragi Kocev,
Pan\v{c}e Panov, Tome Eftimov
- Abstract summary: Multi-label classification (MLC) is an ML task of predictive modeling in which a data instance can simultaneously belong to multiple classes.
Several MLC algorithms have been proposed in the literature, resulting in a meta-optimization problem.
We investigate in this work the quality of an automated approach that uses characteristics of the datasets.
- Score: 6.442438468509492
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-label classification (MLC) is an ML task of predictive modeling in
which a data instance can simultaneously belong to multiple classes. MLC is
increasingly gaining interest in different application domains such as text
mining, computer vision, and bioinformatics. Several MLC algorithms have been
proposed in the literature, resulting in a meta-optimization problem that the
user needs to address: which MLC approach to select for a given dataset? To
address this algorithm selection problem, we investigate in this work the
quality of an automated approach that uses characteristics of the datasets -
so-called features - and a trained algorithm selector to choose which algorithm
to apply for a given task. For our empirical evaluation, we use a portfolio of
38 datasets. We consider eight MLC algorithms, whose quality we evaluate using
six different performance metrics. We show that our automated algorithm
selector outperforms any of the single MLC algorithms, and this is for all
evaluated performance measures. Our selection approach is explainable, a
characteristic that we exploit to investigate which meta-features have the
largest influence on the decisions made by the algorithm selector. Finally, we
also quantify the importance of the most significant meta-features for various
domains.
Related papers
- A Survey of Meta-features Used for Automated Selection of Algorithms for Black-box Single-objective Continuous Optimization [4.173197621837912]
We conduct an overview of the key contributions to algorithm selection in the field of single-objective continuous black-box optimization.
We study machine learning models for automated algorithm selection, configuration, and performance prediction.
arXiv Detail & Related papers (2024-06-08T11:11:14Z) - Multi-objective Binary Coordinate Search for Feature Selection [0.24578723416255746]
We propose the binary multi-objective coordinate search (MOCS) algorithm to solve large-scale feature selection problems.
Results indicate the significant superiority of our method over NSGA-II, on five real-world large-scale datasets.
arXiv Detail & Related papers (2024-02-20T00:50:26Z) - Large Language Model-Enhanced Algorithm Selection: Towards Comprehensive Algorithm Representation [27.378185644892984]
This paper introduces Large Language Models (LLMs) into algorithm selection for the first time.
LLMs not only captures the structural and semantic aspects of the algorithm, but also demonstrates contextual awareness and library function understanding.
The selected algorithm is determined by the matching degree between a given problem and different algorithms.
arXiv Detail & Related papers (2023-11-22T06:23:18Z) - Towards Automated Imbalanced Learning with Deep Hierarchical
Reinforcement Learning [57.163525407022966]
Imbalanced learning is a fundamental challenge in data mining, where there is a disproportionate ratio of training samples in each class.
Over-sampling is an effective technique to tackle imbalanced learning through generating synthetic samples for the minority class.
We propose AutoSMOTE, an automated over-sampling algorithm that can jointly optimize different levels of decisions.
arXiv Detail & Related papers (2022-08-26T04:28:01Z) - Fair Feature Subset Selection using Multiobjective Genetic Algorithm [0.0]
We present a feature subset selection approach that improves both fairness and accuracy objectives.
We use statistical disparity as a fairness metric and F1-Score as a metric for model performance.
Our experiments on the most commonly used fairness benchmark datasets show that using the evolutionary algorithm we can effectively explore the trade-off between fairness and accuracy.
arXiv Detail & Related papers (2022-04-30T22:51:19Z) - Compactness Score: A Fast Filter Method for Unsupervised Feature
Selection [66.84571085643928]
We propose a fast unsupervised feature selection method, named as, Compactness Score (CSUFS) to select desired features.
Our proposed algorithm seems to be more accurate and efficient compared with existing algorithms.
arXiv Detail & Related papers (2022-01-31T13:01:37Z) - Algorithm Selection on a Meta Level [58.720142291102135]
We introduce the problem of meta algorithm selection, which essentially asks for the best way to combine a given set of algorithm selectors.
We present a general methodological framework for meta algorithm selection as well as several concrete learning methods as instantiations of this framework.
arXiv Detail & Related papers (2021-07-20T11:23:21Z) - Memory-Based Optimization Methods for Model-Agnostic Meta-Learning and
Personalized Federated Learning [56.17603785248675]
Model-agnostic meta-learning (MAML) has become a popular research area.
Existing MAML algorithms rely on the episode' idea by sampling a few tasks and data points to update the meta-model at each iteration.
This paper proposes memory-based algorithms for MAML that converge with vanishing error.
arXiv Detail & Related papers (2021-06-09T08:47:58Z) - Towards Meta-Algorithm Selection [78.13985819417974]
Instance-specific algorithm selection (AS) deals with the automatic selection of an algorithm from a fixed set of candidates.
We show that meta-algorithm-selection can indeed prove beneficial in some cases.
arXiv Detail & Related papers (2020-11-17T17:27:33Z) - Automatic selection of clustering algorithms using supervised graph
embedding [14.853602181549967]
MARCO-GE is a novel meta-learning approach for the automated recommendation of clustering algorithms.
It trains a ranking meta-model capable of accurately recommending top-performing algorithms for a new dataset and clustering evaluation measure.
arXiv Detail & Related papers (2020-11-16T19:13:20Z) - Extreme Algorithm Selection With Dyadic Feature Representation [78.13985819417974]
We propose the setting of extreme algorithm selection (XAS) where we consider fixed sets of thousands of candidate algorithms.
We assess the applicability of state-of-the-art AS techniques to the XAS setting and propose approaches leveraging a dyadic feature representation.
arXiv Detail & Related papers (2020-01-29T09:40:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.