DC and SA: Robust and Efficient Hyperparameter Optimization of
Multi-subnetwork Deep Learning Models
- URL: http://arxiv.org/abs/2202.11841v1
- Date: Thu, 24 Feb 2022 00:29:05 GMT
- Title: DC and SA: Robust and Efficient Hyperparameter Optimization of
Multi-subnetwork Deep Learning Models
- Authors: Alex H. Treacher and Albert Montillo
- Abstract summary: We present two novel strategies for optimization of deep learning models with a modular architecture constructed of multipleworks.
Our approaches show an increased optimization efficiency of up to 23.62x, and a final performance boost of up to 3.5% accuracy for classification and 4.4 MSE for regression.
- Score: 0.974672460306765
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: We present two novel hyperparameter optimization strategies for optimization
of deep learning models with a modular architecture constructed of multiple
subnetworks. As complex networks with multiple subnetworks become more
frequently applied in machine learning, hyperparameter optimization methods are
required to efficiently optimize their hyperparameters. Existing hyperparameter
searches are general, and can be used to optimize such networks, however, by
exploiting the multi-subnetwork architecture, these searches can be sped up
substantially. The proposed methods offer faster convergence to a
better-performing final model. To demonstrate this, we propose 2 independent
approaches to enhance these prior algorithms: 1) a divide-and-conquer approach,
in which the best subnetworks of top-performing models are combined, allowing
for more rapid sampling of the hyperparameter search space. 2) A subnetwork
adaptive approach that distributes computational resources based on the
importance of each subnetwork, allowing more intelligent resource allocation.
These approaches can be flexibily applied to many hyperparameter optimization
algorithms. To illustrate this, we combine our approaches with the
commonly-used Bayesian optimization method. Our approaches are then tested
against both synthetic examples and real-world examples and applied to multiple
network types including convolutional neural networks and dense feed forward
neural networks. Our approaches show an increased optimization efficiency of up
to 23.62x, and a final performance boost of up to 3.5% accuracy for
classification and 4.4 MSE for regression, when compared to comparable BO
approach.
Related papers
- Edge-Efficient Deep Learning Models for Automatic Modulation Classification: A Performance Analysis [0.7428236410246183]
We investigate optimized convolutional neural networks (CNNs) developed for automatic modulation classification (AMC) of wireless signals.
We propose optimized models with the combinations of these techniques to fuse the complementary optimization benefits.
The experimental results show that the proposed individual and combined optimization techniques are highly effective for developing models with significantly less complexity.
arXiv Detail & Related papers (2024-04-11T06:08:23Z) - SequentialAttention++ for Block Sparsification: Differentiable Pruning
Meets Combinatorial Optimization [24.55623897747344]
Neural network pruning is a key technique towards engineering large yet scalable, interpretable, generalizable models.
We show how many existing differentiable pruning techniques can be understood as non regularization for group sparse optimization.
We propose SequentialAttention++, which advances state the art in large-scale neural network block-wise pruning tasks on the ImageNet and Criteo datasets.
arXiv Detail & Related papers (2024-02-27T21:42:18Z) - Principled Architecture-aware Scaling of Hyperparameters [69.98414153320894]
Training a high-quality deep neural network requires choosing suitable hyperparameters, which is a non-trivial and expensive process.
In this work, we precisely characterize the dependence of initializations and maximal learning rates on the network architecture.
We demonstrate that network rankings can be easily changed by better training networks in benchmarks.
arXiv Detail & Related papers (2024-02-27T11:52:49Z) - Federated Multi-Level Optimization over Decentralized Networks [55.776919718214224]
We study the problem of distributed multi-level optimization over a network, where agents can only communicate with their immediate neighbors.
We propose a novel gossip-based distributed multi-level optimization algorithm that enables networked agents to solve optimization problems at different levels in a single timescale.
Our algorithm achieves optimal sample complexity, scaling linearly with the network size, and demonstrates state-of-the-art performance on various applications.
arXiv Detail & Related papers (2023-10-10T00:21:10Z) - Bidirectional Looking with A Novel Double Exponential Moving Average to
Adaptive and Non-adaptive Momentum Optimizers [109.52244418498974]
We propose a novel textscAdmeta (textbfADouble exponential textbfMov averagtextbfE textbfAdaptive and non-adaptive momentum) framework.
We provide two implementations, textscAdmetaR and textscAdmetaS, the former based on RAdam and the latter based on SGDM.
arXiv Detail & Related papers (2023-07-02T18:16:06Z) - A Survey on Multi-Objective based Parameter Optimization for Deep
Learning [1.3223682837381137]
We focus on exploring the effectiveness of multi-objective optimization strategies for parameter optimization in conjunction with deep neural networks.
The two methods are combined to provide valuable insights into the generation of predictions and analysis in multiple applications.
arXiv Detail & Related papers (2023-05-17T07:48:54Z) - Application of Monte Carlo Stochastic Optimization (MOST) to Deep
Learning [0.0]
In this paper, we apply the Monte Carlo optimization (MOST) proposed by the authors to a deep learning of XOR gate.
As a result, it was confirmed that it converged faster than the existing method.
arXiv Detail & Related papers (2021-09-02T05:52:26Z) - A self-adapting super-resolution structures framework for automatic
design of GAN [15.351639834230383]
We introduce a new super-resolution image reconstruction generative adversarial network framework.
We use a Bayesian optimization method used to optimize the hyper parameters of the generator and discriminator.
Our method adopts Bayesian optimization as a optimization policy of GAN in our model.
arXiv Detail & Related papers (2021-06-10T19:11:29Z) - Manifold Regularized Dynamic Network Pruning [102.24146031250034]
This paper proposes a new paradigm that dynamically removes redundant filters by embedding the manifold information of all instances into the space of pruned networks.
The effectiveness of the proposed method is verified on several benchmarks, which shows better performance in terms of both accuracy and computational cost.
arXiv Detail & Related papers (2021-03-10T03:59:03Z) - Online hyperparameter optimization by real-time recurrent learning [57.01871583756586]
Our framework takes advantage of the analogy between hyperparameter optimization and parameter learning in neural networks (RNNs)
It adapts a well-studied family of online learning algorithms for RNNs to tune hyperparameters and network parameters simultaneously.
This procedure yields systematically better generalization performance compared to standard methods, at a fraction of wallclock time.
arXiv Detail & Related papers (2021-02-15T19:36:18Z) - Communication-Efficient Distributed Stochastic AUC Maximization with
Deep Neural Networks [50.42141893913188]
We study a distributed variable for large-scale AUC for a neural network as with a deep neural network.
Our model requires a much less number of communication rounds and still a number of communication rounds in theory.
Our experiments on several datasets show the effectiveness of our theory and also confirm our theory.
arXiv Detail & Related papers (2020-05-05T18:08:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.