Generating Efficient DNN-Ensembles with Evolutionary Computation
- URL: http://arxiv.org/abs/2009.08698v2
- Date: Mon, 3 May 2021 12:59:02 GMT
- Title: Generating Efficient DNN-Ensembles with Evolutionary Computation
- Authors: Marc Ortiz, Florian Scheidegger, Marc Casas, Cristiano Malossi, Eduard
Ayguad\'e
- Abstract summary: We leverage ensemble learning as a tool for the creation of faster, smaller, and more accurate deep learning models.
We run EARN on 10 image classification datasets with an initial pool of 32 state-of-the-art DCNN on both CPU and GPU platforms.
We generate models with speedups up to $7.60times$, reductions of parameters by $10times$, or increases in accuracy up to $6.01%$ regarding the best DNN in the pool.
- Score: 3.28217012194635
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this work, we leverage ensemble learning as a tool for the creation of
faster, smaller, and more accurate deep learning models. We demonstrate that we
can jointly optimize for accuracy, inference time, and the number of parameters
by combining DNN classifiers. To achieve this, we combine multiple ensemble
strategies: bagging, boosting, and an ordered chain of classifiers. To reduce
the number of DNN ensemble evaluations during the search, we propose EARN, an
evolutionary approach that optimizes the ensemble according to three objectives
regarding the constraints specified by the user. We run EARN on 10 image
classification datasets with an initial pool of 32 state-of-the-art DCNN on
both CPU and GPU platforms, and we generate models with speedups up to
$7.60\times$, reductions of parameters by $10\times$, or increases in accuracy
up to $6.01\%$ regarding the best DNN in the pool. In addition, our method
generates models that are $5.6\times$ faster than the state-of-the-art methods
for automatic model generation.
Related papers
- Decoding-Time Language Model Alignment with Multiple Objectives [116.42095026960598]
Existing methods primarily focus on optimizing LMs for a single reward function, limiting their adaptability to varied objectives.
Here, we propose $textbfmulti-objective decoding (MOD)$, a decoding-time algorithm that outputs the next token from a linear combination of predictions.
We show why existing approaches can be sub-optimal even in natural settings and obtain optimality guarantees for our method.
arXiv Detail & Related papers (2024-06-27T02:46:30Z) - Multivector Neurons: Better and Faster O(n)-Equivariant Clifford Graph Neural Networks [17.716680490388306]
In this work, we test a few novel message passing graph neural networks (GNNs) based on Clifford multivectors.
We push the state-of-the-art error on the N-body dataset to 0.0035; an 8% improvement over recent methods.
arXiv Detail & Related papers (2024-06-06T13:17:44Z) - ApproxDARTS: Differentiable Neural Architecture Search with Approximate Multipliers [0.24578723416255746]
We present ApproxDARTS, a neural architecture search (NAS) method enabling the popular differentiable neural architecture search method called DARTS to exploit approximate multipliers.
We show that the ApproxDARTS is able to perform a complete architecture search within less than $10$ GPU hours and produce competitive convolutional neural networks (CNN) containing approximate multipliers in convolutional layers.
arXiv Detail & Related papers (2024-04-08T09:54:57Z) - ANN vs SNN: A case study for Neural Decoding in Implantable
Brain-Machine Interfaces [0.7904805552920349]
In this work, we compare different neural networks (NN) for motor decoding in terms of accuracy and implementation cost.
We further show that combining traditional signal processing techniques with machine learning ones deliver surprisingly good performance even with simple NNs.
arXiv Detail & Related papers (2023-12-26T05:40:39Z) - Towards Hyperparameter-Agnostic DNN Training via Dynamical System
Insights [4.513581513983453]
We present a first-order optimization method specialized for deep neural networks (DNNs), ECCO-DNN.
This method models the optimization variable trajectory as a dynamical system and develops a discretization algorithm that adaptively selects step sizes based on the trajectory's shape.
arXiv Detail & Related papers (2023-10-21T03:45:13Z) - Sub-bit Neural Networks: Learning to Compress and Accelerate Binary
Neural Networks [72.81092567651395]
Sub-bit Neural Networks (SNNs) are a new type of binary quantization design tailored to compress and accelerate BNNs.
SNNs are trained with a kernel-aware optimization framework, which exploits binary quantization in the fine-grained convolutional kernel space.
Experiments on visual recognition benchmarks and the hardware deployment on FPGA validate the great potentials of SNNs.
arXiv Detail & Related papers (2021-10-18T11:30:29Z) - ANNETTE: Accurate Neural Network Execution Time Estimation with Stacked
Models [56.21470608621633]
We propose a time estimation framework to decouple the architectural search from the target hardware.
The proposed methodology extracts a set of models from micro- kernel and multi-layer benchmarks and generates a stacked model for mapping and network execution time estimation.
We compare estimation accuracy and fidelity of the generated mixed models, statistical models with the roofline model, and a refined roofline model for evaluation.
arXiv Detail & Related papers (2021-05-07T11:39:05Z) - Binarizing MobileNet via Evolution-based Searching [66.94247681870125]
We propose a use of evolutionary search to facilitate the construction and training scheme when binarizing MobileNet.
Inspired by one-shot architecture search frameworks, we manipulate the idea of group convolution to design efficient 1-Bit Convolutional Neural Networks (CNNs)
Our objective is to come up with a tiny yet efficient binary neural architecture by exploring the best candidates of the group convolution.
arXiv Detail & Related papers (2020-05-13T13:25:51Z) - Self-Directed Online Machine Learning for Topology Optimization [58.920693413667216]
Self-directed Online Learning Optimization integrates Deep Neural Network (DNN) with Finite Element Method (FEM) calculations.
Our algorithm was tested by four types of problems including compliance minimization, fluid-structure optimization, heat transfer enhancement and truss optimization.
It reduced the computational time by 2 5 orders of magnitude compared with directly using methods, and outperformed all state-of-the-art algorithms tested in our experiments.
arXiv Detail & Related papers (2020-02-04T20:00:28Z) - Model Fusion via Optimal Transport [64.13185244219353]
We present a layer-wise model fusion algorithm for neural networks.
We show that this can successfully yield "one-shot" knowledge transfer between neural networks trained on heterogeneous non-i.i.d. data.
arXiv Detail & Related papers (2019-10-12T22:07:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.