EvoPruneDeepTL: An Evolutionary Pruning Model for Transfer Learning
based Deep Neural Networks
- URL: http://arxiv.org/abs/2202.03844v3
- Date: Mon, 5 Feb 2024 12:39:00 GMT
- Title: EvoPruneDeepTL: An Evolutionary Pruning Model for Transfer Learning
based Deep Neural Networks
- Authors: Javier Poyatos, Daniel Molina, Aritz. D. Martinez, Javier Del Ser,
Francisco Herrera
- Abstract summary: We propose an evolutionary pruning model for Transfer Learning based Deep Neural Networks.
EvoPruneDeepTL replaces the last fully-connected layers with sparse layers optimized by a genetic algorithm.
Results show the contribution of EvoPruneDeepTL and feature selection to the overall computational efficiency of the network.
- Score: 15.29595828816055
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In recent years, Deep Learning models have shown a great performance in
complex optimization problems. They generally require large training datasets,
which is a limitation in most practical cases. Transfer learning allows
importing the first layers of a pre-trained architecture and connecting them to
fully-connected layers to adapt them to a new problem. Consequently, the
configuration of the these layers becomes crucial for the performance of the
model. Unfortunately, the optimization of these models is usually a
computationally demanding task. One strategy to optimize Deep Learning models
is the pruning scheme. Pruning methods are focused on reducing the complexity
of the network, assuming an expected performance penalty of the model once
pruned. However, the pruning could potentially be used to improve the
performance, using an optimization algorithm to identify and eventually remove
unnecessary connections among neurons. This work proposes EvoPruneDeepTL, an
evolutionary pruning model for Transfer Learning based Deep Neural Networks
which replaces the last fully-connected layers with sparse layers optimized by
a genetic algorithm. Depending on its solution encoding strategy, our proposed
model can either perform optimized pruning or feature selection over the
densely connected part of the neural network. We carry out different
experiments with several datasets to assess the benefits of our proposal.
Results show the contribution of EvoPruneDeepTL and feature selection to the
overall computational efficiency of the network as a result of the optimization
process. In particular, the accuracy is improved, reducing at the same time the
number of active neurons in the final layers.
Related papers
- Optimizing Dense Feed-Forward Neural Networks [0.0]
We propose a novel feed-forward neural network constructing method based on pruning and transfer learning.
Our approach can compress the number of parameters by more than 70%.
We also evaluate the transfer learning level comparing the refined model and the original one training from scratch a neural network.
arXiv Detail & Related papers (2023-12-16T23:23:16Z) - Explicit Foundation Model Optimization with Self-Attentive Feed-Forward
Neural Units [4.807347156077897]
Iterative approximation methods using backpropagation enable the optimization of neural networks, but they remain computationally expensive when used at scale.
This paper presents an efficient alternative for optimizing neural networks that reduces the costs of scaling neural networks and provides high-efficiency optimizations for low-resource applications.
arXiv Detail & Related papers (2023-11-13T17:55:07Z) - Globally Optimal Training of Neural Networks with Threshold Activation
Functions [63.03759813952481]
We study weight decay regularized training problems of deep neural networks with threshold activations.
We derive a simplified convex optimization formulation when the dataset can be shattered at a certain layer of the network.
arXiv Detail & Related papers (2023-03-06T18:59:13Z) - Multiobjective Evolutionary Pruning of Deep Neural Networks with
Transfer Learning for improving their Performance and Robustness [15.29595828816055]
This work proposes MO-EvoPruneDeepTL, a multi-objective evolutionary pruning algorithm.
We use Transfer Learning to adapt the last layers of Deep Neural Networks, by replacing them with sparse layers evolved by a genetic algorithm.
Experiments show that our proposal achieves promising results in all the objectives, and direct relation are presented.
arXiv Detail & Related papers (2023-02-20T19:33:38Z) - Joint inference and input optimization in equilibrium networks [68.63726855991052]
deep equilibrium model is a class of models that foregoes traditional network depth and instead computes the output of a network by finding the fixed point of a single nonlinear layer.
We show that there is a natural synergy between these two settings.
We demonstrate this strategy on various tasks such as training generative models while optimizing over latent codes, training models for inverse problems like denoising and inpainting, adversarial training and gradient based meta-learning.
arXiv Detail & Related papers (2021-11-25T19:59:33Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z) - Non-Gradient Manifold Neural Network [79.44066256794187]
Deep neural network (DNN) generally takes thousands of iterations to optimize via gradient descent.
We propose a novel manifold neural network based on non-gradient optimization.
arXiv Detail & Related papers (2021-06-15T06:39:13Z) - Analytically Tractable Inference in Deep Neural Networks [0.0]
Tractable Approximate Inference (TAGI) algorithm was shown to be a viable and scalable alternative to backpropagation for shallow fully-connected neural networks.
We are demonstrating how TAGI matches or exceeds the performance of backpropagation, for training classic deep neural network architectures.
arXiv Detail & Related papers (2021-03-09T14:51:34Z) - Large-Scale Gradient-Free Deep Learning with Recursive Local
Representation Alignment [84.57874289554839]
Training deep neural networks on large-scale datasets requires significant hardware resources.
Backpropagation, the workhorse for training these networks, is an inherently sequential process that is difficult to parallelize.
We propose a neuro-biologically-plausible alternative to backprop that can be used to train deep networks.
arXiv Detail & Related papers (2020-02-10T16:20:02Z) - MSE-Optimal Neural Network Initialization via Layer Fusion [68.72356718879428]
Deep neural networks achieve state-of-the-art performance for a range of classification and inference tasks.
The use of gradient combined nonvolutionity renders learning susceptible to novel problems.
We propose fusing neighboring layers of deeper networks that are trained with random variables.
arXiv Detail & Related papers (2020-01-28T18:25:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.