Automatic Componentwise Boosting: An Interpretable AutoML System
- URL: http://arxiv.org/abs/2109.05583v1
- Date: Sun, 12 Sep 2021 18:34:33 GMT
- Title: Automatic Componentwise Boosting: An Interpretable AutoML System
- Authors: Stefan Coors and Daniel Schalk and Bernd Bischl and David R\"ugamer
- Abstract summary: We propose an AutoML system that constructs an interpretable additive model that can be fitted using a highly scalable componentwise boosting algorithm.
Our system provides tools for easy model interpretation such as visualizing partial effects and pairwise interactions.
Despite its restriction to an interpretable model space, our system is competitive in terms of predictive performance on most data sets.
- Score: 1.1709030738577393
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In practice, machine learning (ML) workflows require various different steps,
from data preprocessing, missing value imputation, model selection, to model
tuning as well as model evaluation. Many of these steps rely on human ML
experts. AutoML - the field of automating these ML pipelines - tries to help
practitioners to apply ML off-the-shelf without any expert knowledge. Most
modern AutoML systems like auto-sklearn, H20-AutoML or TPOT aim for high
predictive performance, thereby generating ensembles that consist almost
exclusively of black-box models. This, in turn, makes the interpretation for
the layperson more intricate and adds another layer of opacity for users. We
propose an AutoML system that constructs an interpretable additive model that
can be fitted using a highly scalable componentwise boosting algorithm. Our
system provides tools for easy model interpretation such as visualizing partial
effects and pairwise interactions, allows for a straightforward calculation of
feature importance, and gives insights into the required model complexity to
fit the given task. We introduce the general framework and outline its
implementation autocompboost. To demonstrate the frameworks efficacy, we
compare autocompboost to other existing systems based on the OpenML
AutoML-Benchmark. Despite its restriction to an interpretable model space, our
system is competitive in terms of predictive performance on most data sets
while being more user-friendly and transparent.
Related papers
- Matchmaker: Self-Improving Large Language Model Programs for Schema Matching [60.23571456538149]
We propose a compositional language model program for schema matching, comprised of candidate generation, refinement and confidence scoring.
Matchmaker self-improves in a zero-shot manner without the need for labeled demonstrations.
Empirically, we demonstrate on real-world medical schema matching benchmarks that Matchmaker outperforms previous ML-based approaches.
arXiv Detail & Related papers (2024-10-31T16:34:03Z) - UniAutoML: A Human-Centered Framework for Unified Discriminative and Generative AutoML with Large Language Models [5.725785427377439]
We introduce UniAutoML, a human-centered AutoML framework that unifies AutoML for both discriminative and generative tasks.
The human-centered design of UniAutoML innovatively features a conversational user interface (CUI) that facilitates natural language interactions.
This design enhances transparency and user control throughout the AutoML training process, allowing users to seamlessly break down or modify the model being trained.
arXiv Detail & Related papers (2024-10-09T17:33:15Z) - AutoML-Agent: A Multi-Agent LLM Framework for Full-Pipeline AutoML [56.565200973244146]
Automated machine learning (AutoML) accelerates AI development by automating tasks in the development pipeline.
Recent works have started exploiting large language models (LLM) to lessen such burden.
This paper proposes AutoML-Agent, a novel multi-agent framework tailored for full-pipeline AutoML.
arXiv Detail & Related papers (2024-10-03T20:01:09Z) - Verbalized Machine Learning: Revisiting Machine Learning with Language Models [63.10391314749408]
We introduce the framework of verbalized machine learning (VML)
VML constrains the parameter space to be human-interpretable natural language.
We empirically verify the effectiveness of VML, and hope that VML can serve as a stepping stone to stronger interpretability.
arXiv Detail & Related papers (2024-06-06T17:59:56Z) - Position: A Call to Action for a Human-Centered AutoML Paradigm [83.78883610871867]
Automated machine learning (AutoML) was formed around the fundamental objectives of automatically and efficiently configuring machine learning (ML)
We argue that a key to unlocking AutoML's full potential lies in addressing the currently underexplored aspect of user interaction with AutoML systems.
arXiv Detail & Related papers (2024-06-05T15:05:24Z) - Deciphering AutoML Ensembles: cattleia's Assistance in Decision-Making [0.0]
Cattleia is an application that deciphers the ensembles for regression, multiclass, and binary classification tasks.
It works with models built by three AutoML packages: auto-sklearn, AutoGluon, and FLAML.
arXiv Detail & Related papers (2024-03-19T11:56:21Z) - Multi-modal Auto-regressive Modeling via Visual Words [96.25078866446053]
We propose the concept of visual tokens, which maps the visual features to probability distributions over Large Multi-modal Models' vocabulary.
We further explore the distribution of visual features in the semantic space within LMM and the possibility of using text embeddings to represent visual information.
arXiv Detail & Related papers (2024-03-12T14:58:52Z) - The Devil is in the Errors: Leveraging Large Language Models for
Fine-grained Machine Translation Evaluation [93.01964988474755]
AutoMQM is a prompting technique which asks large language models to identify and categorize errors in translations.
We study the impact of labeled data through in-context learning and finetuning.
We then evaluate AutoMQM with PaLM-2 models, and we find that it improves performance compared to just prompting for scores.
arXiv Detail & Related papers (2023-08-14T17:17:21Z) - Model LineUpper: Supporting Interactive Model Comparison at Multiple
Levels for AutoML [29.04776652873194]
In current AutoML systems, selection is supported only by performance metrics.
We develop tool to support interactive model comparison for AutoML by integrating multiple Explainable AI (XAI) and visualization techniques.
arXiv Detail & Related papers (2021-04-09T14:06:13Z) - Interpret-able feedback for AutoML systems [5.5524559605452595]
Automated machine learning (AutoML) systems aim to enable training machine learning (ML) models for non-ML experts.
A shortcoming of these systems is that when they fail to produce a model with high accuracy, the user has no path to improve the model.
We introduce an interpretable data feedback solution for AutoML.
arXiv Detail & Related papers (2021-02-22T18:54:26Z) - Robusta: Robust AutoML for Feature Selection via Reinforcement Learning [24.24652530951966]
We propose the first robust AutoML framework, Robusta--based on reinforcement learning (RL)
We show that the framework is able to improve the model robustness by up to 22% while maintaining competitive accuracy on benign samples.
arXiv Detail & Related papers (2021-01-15T03:12:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.