On tuning deep learning models: a data mining perspective
- URL: http://arxiv.org/abs/2011.09857v1
- Date: Thu, 19 Nov 2020 14:40:42 GMT
- Title: On tuning deep learning models: a data mining perspective
- Authors: M.M. Ozturk
- Abstract summary: Four types of deep learning algorithms are investigated in terms of tuning and data mining perspective.
The number of features has not contributed to the decline in the accuracy of deep learning algorithms.
A uniform distribution is much more crucial to reach reliable results in terms of data mining.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning algorithms vary depending on the underlying connection
mechanism of nodes of them. They have various hyperparameters that are either
set via specific algorithms or randomly chosen. Meanwhile, hyperparameters of
deep learning algorithms have the potential to help enhance the performance of
the machine learning tasks. In this paper, a tuning guideline is provided for
researchers who cope with issues originated from hyperparameters of deep
learning models. To that end, four types of deep learning algorithms are
investigated in terms of tuning and data mining perspective. Further, common
search methods of hyperparameters are evaluated on four deep learning
algorithms. Normalization helps increase the performance of classification,
according to the results of this study. The number of features has not
contributed to the decline in the accuracy of deep learning algorithms. Even
though high sparsity results in low accuracy, a uniform distribution is much
more crucial to reach reliable results in terms of data mining.
Related papers
- Information plane and compression-gnostic feedback in quantum machine learning [0.0]
The information plane has been proposed as an analytical tool for studying the learning dynamics of neural networks.
We study how the insight on how much the model compresses the input data can be used to improve a learning algorithm.
We benchmark the proposed learning algorithms on several classification and regression tasks using variational quantum circuits.
arXiv Detail & Related papers (2024-11-04T17:38:46Z) - A Comparative Study of Hyperparameter Tuning Methods [0.0]
Tree-structured Parzen Estimator (TPE), Genetic Search, and Random Search are evaluated across regression and classification tasks.
Random Search excelled in regression tasks, while TPE was more effective for classification tasks.
arXiv Detail & Related papers (2024-08-29T10:35:07Z) - Robustness of Algorithms for Causal Structure Learning to Hyperparameter
Choice [2.3020018305241337]
Hyper parameter tuning can make the difference between state-of-the-art and poor prediction performance for any algorithm.
We investigate the influence of hyper parameter selection on causal structure learning tasks.
arXiv Detail & Related papers (2023-10-27T15:34:08Z) - Unsupervised Learning of Initialization in Deep Neural Networks via
Maximum Mean Discrepancy [74.34895342081407]
We propose an unsupervised algorithm to find good initialization for input data.
We first notice that each parameter configuration in the parameter space corresponds to one particular downstream task of d-way classification.
We then conjecture that the success of learning is directly related to how diverse downstream tasks are in the vicinity of the initial parameters.
arXiv Detail & Related papers (2023-02-08T23:23:28Z) - Deep networks for system identification: a Survey [56.34005280792013]
System identification learns mathematical descriptions of dynamic systems from input-output data.
Main aim of the identified model is to predict new data from previous observations.
We discuss architectures commonly adopted in the literature, like feedforward, convolutional, and recurrent networks.
arXiv Detail & Related papers (2023-01-30T12:38:31Z) - Improved Algorithms for Neural Active Learning [74.89097665112621]
We improve the theoretical and empirical performance of neural-network(NN)-based active learning algorithms for the non-parametric streaming setting.
We introduce two regret metrics by minimizing the population loss that are more suitable in active learning than the one used in state-of-the-art (SOTA) related work.
arXiv Detail & Related papers (2022-10-02T05:03:38Z) - Exploration of Various Deep Learning Models for Increased Accuracy in
Automatic Polyp Detection [62.997667081978825]
This paper explores deep learning models and algorithms that results in highest accuracy in detecting polyp on colonoscopy images.
Previous studies implemented deep learning using convolution neural network (CNN)
arXiv Detail & Related papers (2022-03-04T04:03:41Z) - A Pragmatic Look at Deep Imitation Learning [0.3626013617212666]
We re-implement 6 different adversarial imitation learning algorithms.
We evaluate them on a widely-used expert trajectory dataset.
GAIL consistently performs well across a range of sample sizes.
arXiv Detail & Related papers (2021-08-04T06:33:10Z) - Evolving Reinforcement Learning Algorithms [186.62294652057062]
We propose a method for meta-learning reinforcement learning algorithms.
The learned algorithms are domain-agnostic and can generalize to new environments not seen during training.
We highlight two learned algorithms which obtain good generalization performance over other classical control tasks, gridworld type tasks, and Atari games.
arXiv Detail & Related papers (2021-01-08T18:55:07Z) - Towards Optimally Efficient Tree Search with Deep Learning [76.64632985696237]
This paper investigates the classical integer least-squares problem which estimates signals integer from linear models.
The problem is NP-hard and often arises in diverse applications such as signal processing, bioinformatics, communications and machine learning.
We propose a general hyper-accelerated tree search (HATS) algorithm by employing a deep neural network to estimate the optimal estimation for the underlying simplified memory-bounded A* algorithm.
arXiv Detail & Related papers (2021-01-07T08:00:02Z) - Enhancing accuracy of deep learning algorithms by training with
low-discrepancy sequences [15.2292571922932]
We propose a deep supervised learning algorithm based on low-discrepancy sequences as the training set.
We demonstrate that the proposed algorithm significantly outperforms standard deep learning algorithms for problems in moderately high dimensions.
arXiv Detail & Related papers (2020-05-26T08:14:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.