Towards a learning-based performance modeling for accelerating Deep
Neural Networks
- URL: http://arxiv.org/abs/2212.05031v1
- Date: Fri, 9 Dec 2022 18:28:07 GMT
- Title: Towards a learning-based performance modeling for accelerating Deep
Neural Networks
- Authors: Damiano Perri, Paolo Sylos Labini, Osvaldo Gervasi, Sergio Tasso,
Flavio Vella
- Abstract summary: We start an investigation of predictive models based on machine learning techniques in order to optimize Convolution Neural Networks (CNNs)
Preliminary experiments on Midgard-based ARM Mali GPU show that our predictive model outperforms all the convolution operators manually selected by the library.
- Score: 1.1549572298362785
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Emerging applications such as Deep Learning are often data-driven, thus
traditional approaches based on auto-tuners are not performance effective
across the wide range of inputs used in practice. In the present paper, we
start an investigation of predictive models based on machine learning
techniques in order to optimize Convolution Neural Networks (CNNs). As a
use-case, we focus on the ARM Compute Library which provides three different
implementations of the convolution operator at different numeric precision.
Starting from a collation of benchmarks, we build and validate models learned
by Decision Tree and naive Bayesian classifier. Preliminary experiments on
Midgard-based ARM Mali GPU show that our predictive model outperforms all the
convolution operators manually selected by the library.
Related papers
- BEND: Bagging Deep Learning Training Based on Efficient Neural Network Diffusion [56.9358325168226]
We propose a Bagging deep learning training algorithm based on Efficient Neural network Diffusion (BEND)
Our approach is simple but effective, first using multiple trained model weights and biases as inputs to train autoencoder and latent diffusion model.
Our proposed BEND algorithm can consistently outperform the mean and median accuracies of both the original trained model and the diffused model.
arXiv Detail & Related papers (2024-03-23T08:40:38Z) - Diffusion-Model-Assisted Supervised Learning of Generative Models for
Density Estimation [10.793646707711442]
We present a framework for training generative models for density estimation.
We use the score-based diffusion model to generate labeled data.
Once the labeled data are generated, we can train a simple fully connected neural network to learn the generative model in the supervised manner.
arXiv Detail & Related papers (2023-10-22T23:56:19Z) - Training dynamic models using early exits for automatic speech
recognition on resource-constrained devices [15.879328412777008]
Early-exit architectures enable the development of dynamic models capable of adapting their size and architecture to varying levels of computational resources and ASR performance demands.
We show that early-exit models trained from scratch not only preserve performance when using fewer encoder layers but also exhibit enhanced task accuracy compared to single-exit or pre-trained models.
Results provide insights into the training dynamics of early-exit architectures for ASR models.
arXiv Detail & Related papers (2023-09-18T07:45:16Z) - HyperImpute: Generalized Iterative Imputation with Automatic Model
Selection [77.86861638371926]
We propose a generalized iterative imputation framework for adaptively and automatically configuring column-wise models.
We provide a concrete implementation with out-of-the-box learners, simulators, and interfaces.
arXiv Detail & Related papers (2022-06-15T19:10:35Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z) - A Near-Optimal Algorithm for Debiasing Trained Machine Learning Models [21.56208997475512]
We present a scalable post-processing algorithm for debiasing trained models, including deep neural networks (DNNs)
We prove to be near-optimal by bounding its excess Bayes risk.
We empirically validate its advantages on standard benchmark datasets.
arXiv Detail & Related papers (2021-06-06T09:45:37Z) - ALT-MAS: A Data-Efficient Framework for Active Testing of Machine
Learning Algorithms [58.684954492439424]
We propose a novel framework to efficiently test a machine learning model using only a small amount of labeled test data.
The idea is to estimate the metrics of interest for a model-under-test using Bayesian neural network (BNN)
arXiv Detail & Related papers (2021-04-11T12:14:04Z) - Few-Shot Named Entity Recognition: A Comprehensive Study [92.40991050806544]
We investigate three schemes to improve the model generalization ability for few-shot settings.
We perform empirical comparisons on 10 public NER datasets with various proportions of labeled data.
We create new state-of-the-art results on both few-shot and training-free settings.
arXiv Detail & Related papers (2020-12-29T23:43:16Z) - On Training and Evaluation of Neural Network Approaches for Model
Predictive Control [9.8918553325509]
This paper is a framework for training and evaluation of Model Predictive Control (MPC) implemented using constrained neural networks.
The motivation is to replace real-time optimization in safety critical feedback control systems with learnt mappings in the form of neural networks with optimization layers.
arXiv Detail & Related papers (2020-05-08T15:37:55Z) - Model Fusion via Optimal Transport [64.13185244219353]
We present a layer-wise model fusion algorithm for neural networks.
We show that this can successfully yield "one-shot" knowledge transfer between neural networks trained on heterogeneous non-i.i.d. data.
arXiv Detail & Related papers (2019-10-12T22:07:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.