Efficient Hyperparameter Optimization in Deep Learning Using a Variable
Length Genetic Algorithm
- URL: http://arxiv.org/abs/2006.12703v1
- Date: Tue, 23 Jun 2020 02:37:14 GMT
- Title: Efficient Hyperparameter Optimization in Deep Learning Using a Variable
Length Genetic Algorithm
- Authors: Xueli Xiao, Ming Yan, Sunitha Basodi, Chunyan Ji, Yi Pan
- Abstract summary: We propose a variable length genetic algorithm (GA) to systematically and automatically tune the hyper parameters of a CNN.
Experimental results show that our algorithm can find good CNN hyper parameters efficiently.
- Score: 10.497188475929299
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Convolutional Neural Networks (CNN) have gained great success in many
artificial intelligence tasks. However, finding a good set of hyperparameters
for a CNN remains a challenging task. It usually takes an expert with deep
knowledge, and trials and errors. Genetic algorithms have been used in
hyperparameter optimizations. However, traditional genetic algorithms with
fixed-length chromosomes may not be a good fit for optimizing deep learning
hyperparameters, because deep learning models have variable number of
hyperparameters depending on the model depth. As the depth increases, the
number of hyperparameters grows exponentially, and searching becomes
exponentially harder. It is important to have an efficient algorithm that can
find a good model in reasonable time. In this article, we propose to use a
variable length genetic algorithm (GA) to systematically and automatically tune
the hyperparameters of a CNN to improve its performance. Experimental results
show that our algorithm can find good CNN hyperparameters efficiently. It is
clear from our experiments that if more time is spent on optimizing the
hyperparameters, better results could be achieved. Theoretically, if we had
unlimited time and CPU power, we could find the optimized hyperparameters and
achieve the best results in the future.
Related papers
- Optimization of Convolutional Neural Network Hyperparameter for Medical Image Diagnosis using Metaheuristic Algorithms: A short Recent Review (2019-2022) [3.3686252536891454]
Convolutional Neural Networks (CNNs) have been successfully utilized in the medical diagnosis of many illnesses.
This study will present work done in recent years on the usage of metaheuristic optimization algorithms in the CNN optimization process.
arXiv Detail & Related papers (2024-12-23T20:13:02Z) - Massively Parallel Genetic Optimization through Asynchronous Propagation
of Populations [50.591267188664666]
Propulate is an evolutionary optimization algorithm and software package for global optimization.
We provide an MPI-based implementation of our algorithm, which features variants of selection, mutation, crossover, and migration.
We find that Propulate is up to three orders of magnitude faster without sacrificing solution accuracy.
arXiv Detail & Related papers (2023-01-20T18:17:34Z) - AUTOMATA: Gradient Based Data Subset Selection for Compute-Efficient
Hyper-parameter Tuning [72.54359545547904]
We propose a gradient-based subset selection framework for hyper- parameter tuning.
We show that using gradient-based data subsets for hyper- parameter tuning achieves significantly faster turnaround times and speedups of 3$times$-30$times$.
arXiv Detail & Related papers (2022-03-15T19:25:01Z) - HyP-ABC: A Novel Automated Hyper-Parameter Tuning Algorithm Using
Evolutionary Optimization [1.6114012813668934]
We propose HyP-ABC, an automatic hybrid hyper-parameter optimization algorithm using the modified artificial bee colony approach.
Compared to the state-of-the-art techniques, HyP-ABC is more efficient and has a limited number of parameters to be tuned.
arXiv Detail & Related papers (2021-09-11T16:45:39Z) - HyperNP: Interactive Visual Exploration of Multidimensional Projection
Hyperparameters [61.354362652006834]
HyperNP is a scalable method that allows for real-time interactive exploration of projection methods by training neural network approximations.
We evaluate the performance of the HyperNP across three datasets in terms of performance and speed.
arXiv Detail & Related papers (2021-06-25T17:28:14Z) - Optimizing Large-Scale Hyperparameters via Automated Learning Algorithm [97.66038345864095]
We propose a new hyperparameter optimization method with zeroth-order hyper-gradients (HOZOG)
Specifically, we first formulate hyperparameter optimization as an A-based constrained optimization problem.
Then, we use the average zeroth-order hyper-gradients to update hyper parameters.
arXiv Detail & Related papers (2021-02-17T21:03:05Z) - Online hyperparameter optimization by real-time recurrent learning [57.01871583756586]
Our framework takes advantage of the analogy between hyperparameter optimization and parameter learning in neural networks (RNNs)
It adapts a well-studied family of online learning algorithms for RNNs to tune hyperparameters and network parameters simultaneously.
This procedure yields systematically better generalization performance compared to standard methods, at a fraction of wallclock time.
arXiv Detail & Related papers (2021-02-15T19:36:18Z) - Automatic Setting of DNN Hyper-Parameters by Mixing Bayesian
Optimization and Tuning Rules [0.6875312133832078]
We build a new algorithm for evaluating and analyzing the results of the network on the training and validation sets.
We use a set of tuning rules to add new hyper-parameters and/or to reduce the hyper- parameter search space to select a better combination.
arXiv Detail & Related papers (2020-06-03T08:53:48Z) - Weighted Random Search for CNN Hyperparameter Optimization [0.0]
We introduce the weighted Random Search (WRS) method, a combination of Random Search (RS) and probabilistic greedy.
The criterion is the classification accuracy achieved within the same number of tested combinations of hyperparameter values.
According to our experiments, the WRS algorithm outperforms the other methods.
arXiv Detail & Related papers (2020-03-30T09:40:14Z) - Self-Directed Online Machine Learning for Topology Optimization [58.920693413667216]
Self-directed Online Learning Optimization integrates Deep Neural Network (DNN) with Finite Element Method (FEM) calculations.
Our algorithm was tested by four types of problems including compliance minimization, fluid-structure optimization, heat transfer enhancement and truss optimization.
It reduced the computational time by 2 5 orders of magnitude compared with directly using methods, and outperformed all state-of-the-art algorithms tested in our experiments.
arXiv Detail & Related papers (2020-02-04T20:00:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.