A Deep Neural Networks ensemble workflow from hyperparameter search to
inference leveraging GPU clusters
- URL: http://arxiv.org/abs/2208.14046v1
- Date: Tue, 30 Aug 2022 08:04:19 GMT
- Title: A Deep Neural Networks ensemble workflow from hyperparameter search to
inference leveraging GPU clusters
- Authors: Pierrick Pochelu, Serge G. Petiton, Bruno Conche
- Abstract summary: AutoML seeks to automatically build ensembles of Deep Neural Networks (DNNs) to achieve qualitative predictions.
We propose a new AutoML to build a larger library of accurate and diverse individual models to then construct ensembles.
New ensemble selection method based on a multi-objective greedy algorithm is proposed to generate accurate ensembles.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Automated Machine Learning with ensembling (or AutoML with ensembling) seeks
to automatically build ensembles of Deep Neural Networks (DNNs) to achieve
qualitative predictions. Ensemble of DNNs are well known to avoid over-fitting
but they are memory and time consuming approaches. Therefore, an ideal AutoML
would produce in one single run time different ensembles regarding accuracy and
inference speed. While previous works on AutoML focus to search for the best
model to maximize its generalization ability, we rather propose a new AutoML to
build a larger library of accurate and diverse individual models to then
construct ensembles. First, our extensive benchmarks show asynchronous
Hyperband is an efficient and robust way to build a large number of diverse
models to combine them. Then, a new ensemble selection method based on a
multi-objective greedy algorithm is proposed to generate accurate ensembles by
controlling their computing cost. Finally, we propose a novel algorithm to
optimize the inference of the DNNs ensemble in a GPU cluster based on
allocation optimization. The produced AutoML with ensemble method shows robust
results on two datasets using efficiently GPU clusters during both the training
phase and the inference phase.
Related papers
- CLAMS: A System for Zero-Shot Model Selection for Clustering [3.7127285734321194]
We propose an AutoML system that enables model selection on clustering problems by leveraging optimal transport-based dataset similarity.
We compare our results against multiple clustering baselines and find that it outperforms all of them, hence demonstrating the utility of similarity-based automated model selection for solving clustering applications.
arXiv Detail & Related papers (2024-07-15T23:50:07Z) - Efficient Automated Deep Learning for Time Series Forecasting [42.47842694670572]
We propose an efficient approach for the joint optimization of neural architecture and hyperparameters of the entire data processing pipeline for time series forecasting.
In contrast to common NAS search spaces, we designed a novel neural architecture search space covering various state-of-the-art architectures.
We empirically study several different budget types enabling efficient multi-fidelity optimization on different forecasting datasets.
arXiv Detail & Related papers (2022-05-11T14:03:25Z) - Mixture Model Auto-Encoders: Deep Clustering through Dictionary Learning [72.9458277424712]
Mixture Model Auto-Encoders (MixMate) is a novel architecture that clusters data by performing inference on a generative model.
We show that MixMate achieves competitive performance compared to state-of-the-art deep clustering algorithms.
arXiv Detail & Related papers (2021-10-10T02:30:31Z) - Hierarchical Clustering using Auto-encoded Compact Representation for
Time-series Analysis [8.660029077292346]
We propose a novel mechanism to identify the clusters combining learned compact representation of time-series, Auto Encoded Compact Sequence (AECS) and hierarchical clustering approach.
Our algorithm exploits Recurrent Neural Network (RNN) based under complete Sequence to Sequence(seq2seq) autoencoder and agglomerative hierarchical clustering.
arXiv Detail & Related papers (2021-01-11T08:03:57Z) - Solving Mixed Integer Programs Using Neural Networks [57.683491412480635]
This paper applies learning to the two key sub-tasks of a MIP solver, generating a high-quality joint variable assignment, and bounding the gap in objective value between that assignment and an optimal one.
Our approach constructs two corresponding neural network-based components, Neural Diving and Neural Branching, to use in a base MIP solver such as SCIP.
We evaluate our approach on six diverse real-world datasets, including two Google production datasets and MIPLIB, by training separate neural networks on each.
arXiv Detail & Related papers (2020-12-23T09:33:11Z) - AgEBO-Tabular: Joint Neural Architecture and Hyperparameter Search with
Autotuned Data-Parallel Training for Tabular Data [11.552769149674544]
Development of high-performing predictive models for large data sets is a challenging task.
Recent automated machine learning (AutoML) is emerging as a promising approach to automate predictive model development.
We have developed AgEBO-Tabular, an approach to combine aging evolution (AgE) and a parallel NAS method that searches over neural architecture space.
arXiv Detail & Related papers (2020-10-30T16:28:48Z) - Fast, Accurate, and Simple Models for Tabular Data via Augmented
Distillation [97.42894942391575]
We propose FAST-DAD to distill arbitrarily complex ensemble predictors into individual models like boosted trees, random forests, and deep networks.
Our individual distilled models are over 10x faster and more accurate than ensemble predictors produced by AutoML tools like H2O/AutoSklearn.
arXiv Detail & Related papers (2020-06-25T09:57:47Z) - Auto-PyTorch Tabular: Multi-Fidelity MetaLearning for Efficient and
Robust AutoDL [53.40030379661183]
Auto-PyTorch is a framework to enable fully automated deep learning (AutoDL)
It combines multi-fidelity optimization with portfolio construction for warmstarting and ensembling of deep neural networks (DNNs)
We show that Auto-PyTorch performs better than several state-of-the-art competitors on average.
arXiv Detail & Related papers (2020-06-24T15:15:17Z) - Stepwise Model Selection for Sequence Prediction via Deep Kernel
Learning [100.83444258562263]
We propose a novel Bayesian optimization (BO) algorithm to tackle the challenge of model selection in this setting.
In order to solve the resulting multiple black-box function optimization problem jointly and efficiently, we exploit potential correlations among black-box functions.
We are the first to formulate the problem of stepwise model selection (SMS) for sequence prediction, and to design and demonstrate an efficient joint-learning algorithm for this purpose.
arXiv Detail & Related papers (2020-01-12T09:42:19Z) - Model Fusion via Optimal Transport [64.13185244219353]
We present a layer-wise model fusion algorithm for neural networks.
We show that this can successfully yield "one-shot" knowledge transfer between neural networks trained on heterogeneous non-i.i.d. data.
arXiv Detail & Related papers (2019-10-12T22:07:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.