To tune or not to tune? An Approach for Recommending Important
Hyperparameters
- URL: http://arxiv.org/abs/2108.13066v1
- Date: Mon, 30 Aug 2021 08:54:58 GMT
- Title: To tune or not to tune? An Approach for Recommending Important
Hyperparameters
- Authors: Mohamadjavad Bahmani, Radwa El Shawi, Nshan Potikyan, Sherif Sakr
- Abstract summary: We consider building the relationship between the performance of the machine learning models and their hyperparameters to discover the trend and gain insights.
Our results enable users to decide whether it is worth conducting a possibly time-consuming tuning strategy.
- Score: 2.121963121603413
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Novel technologies in automated machine learning ease the complexity of
algorithm selection and hyperparameter optimization. Hyperparameters are
important for machine learning models as they significantly influence the
performance of machine learning models. Many optimization techniques have
achieved notable success in hyperparameter tuning and surpassed the performance
of human experts. However, depending on such techniques as blackbox algorithms
can leave machine learning practitioners without insight into the relative
importance of different hyperparameters. In this paper, we consider building
the relationship between the performance of the machine learning models and
their hyperparameters to discover the trend and gain insights, with empirical
results based on six classifiers and 200 datasets. Our results enable users to
decide whether it is worth conducting a possibly time-consuming tuning
strategy, to focus on the most important hyperparameters, and to choose
adequate hyperparameter spaces for tuning. The results of our experiments show
that gradient boosting and Adaboost outperform other classifiers across 200
problems. However, they need tuning to boost their performance. Overall, the
results obtained from this study provide a quantitative basis to focus efforts
toward guided automated hyperparameter optimization and contribute toward the
development of better-automated machine learning frameworks.
Related papers
- Hyperparameter Optimization in Machine Learning [34.356747514732966]
Hyperparameters are configuration variables controlling the behavior of machine learning algorithms.
The choice of their values determine the effectiveness of systems based on these technologies.
We present a unified treatment of hyperparameter optimization, providing the reader with examples and insights into the state-of-the-art.
arXiv Detail & Related papers (2024-10-30T09:39:22Z) - Efficient Hyperparameter Importance Assessment for CNNs [1.7778609937758323]
This paper aims to quantify the importance weights of some hyperparameters in Convolutional Neural Networks (CNNs) with an algorithm called N-RReliefF.
We conduct an extensive study by training over ten thousand CNN models across ten popular image classification datasets.
arXiv Detail & Related papers (2024-10-11T15:47:46Z) - AutoRL Hyperparameter Landscapes [69.15927869840918]
Reinforcement Learning (RL) has shown to be capable of producing impressive results, but its use is limited by the impact of its hyperparameters on performance.
We propose an approach to build and analyze these hyperparameter landscapes not just for one point in time but at multiple points in time throughout training.
This supports the theory that hyperparameters should be dynamically adjusted during training and shows the potential for more insights on AutoRL problems that can be gained through landscape analyses.
arXiv Detail & Related papers (2023-04-05T12:14:41Z) - Hyper-Parameter Auto-Tuning for Sparse Bayesian Learning [72.83293818245978]
We design and learn a neural network (NN)-based auto-tuner for hyper- parameter tuning in sparse Bayesian learning.
We show that considerable improvement in convergence rate and recovery performance can be achieved.
arXiv Detail & Related papers (2022-11-09T12:34:59Z) - AUTOMATA: Gradient Based Data Subset Selection for Compute-Efficient
Hyper-parameter Tuning [72.54359545547904]
We propose a gradient-based subset selection framework for hyper- parameter tuning.
We show that using gradient-based data subsets for hyper- parameter tuning achieves significantly faster turnaround times and speedups of 3$times$-30$times$.
arXiv Detail & Related papers (2022-03-15T19:25:01Z) - Amortized Auto-Tuning: Cost-Efficient Transfer Optimization for
Hyperparameter Recommendation [83.85021205445662]
We propose an instantiation--amortized auto-tuning (AT2) to speed up tuning of machine learning models.
We conduct a thorough analysis of the multi-task multi-fidelity Bayesian optimization framework, which leads to the best instantiation--amortized auto-tuning (AT2)
arXiv Detail & Related papers (2021-06-17T00:01:18Z) - Guided Hyperparameter Tuning Through Visualization and Inference [12.035299005299306]
We present a streamlined visualization system enabling deep learning practitioners to more efficiently explore, tune, and optimize hyper parameters.
A key idea is to directly suggest more optimal hyper parameters using a predictive mechanism.
We evaluate the tool with a user study on deep learning model builders, finding that our participants have little issue adopting the tool and working with it as part of their workflow.
arXiv Detail & Related papers (2021-05-24T19:55:24Z) - Search Algorithms for Automated Hyper-Parameter Tuning [1.2233362977312945]
We develop two automated Hyper- Optimization methods, namely grid search and random search, to assess and improve a previous study's performance.
Experiment results show that applying random search and grid search on machine learning algorithms improves accuracy.
arXiv Detail & Related papers (2021-04-29T22:11:52Z) - Optimizing Large-Scale Hyperparameters via Automated Learning Algorithm [97.66038345864095]
We propose a new hyperparameter optimization method with zeroth-order hyper-gradients (HOZOG)
Specifically, we first formulate hyperparameter optimization as an A-based constrained optimization problem.
Then, we use the average zeroth-order hyper-gradients to update hyper parameters.
arXiv Detail & Related papers (2021-02-17T21:03:05Z) - How much progress have we made in neural network training? A New
Evaluation Protocol for Benchmarking Optimizers [86.36020260204302]
We propose a new benchmarking protocol to evaluate both end-to-end efficiency and data-addition training efficiency.
A human study is conducted to show that our evaluation protocol matches human tuning behavior better than the random search.
We then apply the proposed benchmarking framework to 7s and various tasks, including computer vision, natural language processing, reinforcement learning, and graph mining.
arXiv Detail & Related papers (2020-10-19T21:46:39Z) - On Hyperparameter Optimization of Machine Learning Algorithms: Theory
and Practice [10.350337750192997]
We introduce several state-of-the-art optimization techniques and discuss how to apply them to machine learning algorithms.
This paper will help industrial users, data analysts, and researchers to better develop machine learning models.
arXiv Detail & Related papers (2020-07-30T21:11:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.