Optimization of Convolutional Neural Network Using the Linearly
Decreasing Weight Particle Swarm Optimization
- URL: http://arxiv.org/abs/2001.05670v2
- Date: Thu, 17 Sep 2020 11:49:42 GMT
- Title: Optimization of Convolutional Neural Network Using the Linearly
Decreasing Weight Particle Swarm Optimization
- Authors: T. Serizawa, H. Fujita
- Abstract summary: Convolutional neural network (CNN) is one of the most frequently used deep learning techniques.
In this paper, we propose CNN hyperparameter optimization with linearly decreasing weight particle swarm optimization (LDWPSO)
As a result, when using the MNIST dataset, the baseline CNN is 94.02% at the 5th epoch, compared to 98.95% for LDWPSO CNN, which improves accuracy.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Convolutional neural network (CNN) is one of the most frequently used deep
learning techniques. Various forms of models have been proposed and improved
for learning at CNN. When learning with CNN, it is necessary to determine the
optimal hyperparameters. However, the number of hyperparameters is so large
that it is difficult to do it manually, so much research has been done on
automation. A method that uses metaheuristic algorithms is attracting attention
in research on hyperparameter optimization. Metaheuristic algorithms are
naturally inspired and include evolution strategies, genetic algorithms,
antcolony optimization and particle swarm optimization. In particular, particle
swarm optimization converges faster than genetic algorithms, and various models
have been proposed. In this paper, we propose CNN hyperparameter optimization
with linearly decreasing weight particle swarm optimization (LDWPSO). In the
experiment, the MNIST data set and CIFAR-10 data set, which are often used as
benchmark data sets, are used. By optimizing CNN hyperparameters with LDWPSO,
learning the MNIST and CIFAR-10 datasets, we compare the accuracy with a
standard CNN based on LeNet-5. As a result, when using the MNIST dataset, the
baseline CNN is 94.02% at the 5th epoch, compared to 98.95% for LDWPSO CNN,
which improves accuracy. When using the CIFAR-10 dataset, the Baseline CNN is
28.07% at the 10th epoch, compared to 69.37% for the LDWPSO CNN, which greatly
improves accuracy.
Related papers
- Optimization of Convolutional Neural Network Hyperparameter for Medical Image Diagnosis using Metaheuristic Algorithms: A short Recent Review (2019-2022) [3.3686252536891454]
Convolutional Neural Networks (CNNs) have been successfully utilized in the medical diagnosis of many illnesses.
This study will present work done in recent years on the usage of metaheuristic optimization algorithms in the CNN optimization process.
arXiv Detail & Related papers (2024-12-23T20:13:02Z) - Just How Flexible are Neural Networks in Practice? [89.80474583606242]
It is widely believed that a neural network can fit a training set containing at least as many samples as it has parameters.
In practice, however, we only find solutions via our training procedure, including the gradient and regularizers, limiting flexibility.
arXiv Detail & Related papers (2024-06-17T12:24:45Z) - An Efficient Evolutionary Deep Learning Framework Based on Multi-source
Transfer Learning to Evolve Deep Convolutional Neural Networks [8.40112153818812]
Convolutional neural networks (CNNs) have constantly achieved better performance over years by introducing more complex topology, and enlarging the capacity towards deeper and wider CNNs.
The computational cost is still the bottleneck of automatically designing CNNs.
In this paper, inspired by transfer learning, a new evolutionary computation based framework is proposed to efficiently evolve CNNs.
arXiv Detail & Related papers (2022-12-07T20:22:58Z) - VeLO: Training Versatile Learned Optimizers by Scaling Up [67.90237498659397]
We leverage the same scaling approach behind the success of deep learning to learn versatiles.
We train an ingest for deep learning which is itself a small neural network that ingests and outputs parameter updates.
We open source our learned, meta-training code, the associated train test data, and an extensive benchmark suite with baselines at velo-code.io.
arXiv Detail & Related papers (2022-11-17T18:39:07Z) - Convolution Neural Network Hyperparameter Optimization Using Simplified
Swarm Optimization [2.322689362836168]
Convolutional Neural Network (CNN) is widely used in computer vision.
It is not easy to find a network architecture with better performance.
arXiv Detail & Related papers (2021-03-06T00:23:27Z) - Fixed-point Quantization of Convolutional Neural Networks for Quantized
Inference on Embedded Platforms [0.9954382983583577]
We propose a method to optimally quantize the weights, biases and activations of each layer of a pre-trained CNN.
We find that layer-wise quantization of parameters significantly helps in this process.
arXiv Detail & Related papers (2021-02-03T17:05:55Z) - Kernel Based Progressive Distillation for Adder Neural Networks [71.731127378807]
Adder Neural Networks (ANNs) which only contain additions bring us a new way of developing deep neural networks with low energy consumption.
There is an accuracy drop when replacing all convolution filters by adder filters.
We present a novel method for further improving the performance of ANNs without increasing the trainable parameters.
arXiv Detail & Related papers (2020-09-28T03:29:19Z) - RIFLE: Backpropagation in Depth for Deep Transfer Learning through
Re-Initializing the Fully-connected LayEr [60.07531696857743]
Fine-tuning the deep convolution neural network(CNN) using a pre-trained model helps transfer knowledge learned from larger datasets to the target task.
We propose RIFLE - a strategy that deepens backpropagation in transfer learning settings.
RIFLE brings meaningful updates to the weights of deep CNN layers and improves low-level feature learning.
arXiv Detail & Related papers (2020-07-07T11:27:43Z) - Efficient Hyperparameter Optimization in Deep Learning Using a Variable
Length Genetic Algorithm [10.497188475929299]
We propose a variable length genetic algorithm (GA) to systematically and automatically tune the hyper parameters of a CNN.
Experimental results show that our algorithm can find good CNN hyper parameters efficiently.
arXiv Detail & Related papers (2020-06-23T02:37:14Z) - Self-Directed Online Machine Learning for Topology Optimization [58.920693413667216]
Self-directed Online Learning Optimization integrates Deep Neural Network (DNN) with Finite Element Method (FEM) calculations.
Our algorithm was tested by four types of problems including compliance minimization, fluid-structure optimization, heat transfer enhancement and truss optimization.
It reduced the computational time by 2 5 orders of magnitude compared with directly using methods, and outperformed all state-of-the-art algorithms tested in our experiments.
arXiv Detail & Related papers (2020-02-04T20:00:28Z) - Approximation and Non-parametric Estimation of ResNet-type Convolutional
Neural Networks [52.972605601174955]
We show a ResNet-type CNN can attain the minimax optimal error rates in important function classes.
We derive approximation and estimation error rates of the aformentioned type of CNNs for the Barron and H"older classes.
arXiv Detail & Related papers (2019-03-24T19:42:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.