ImmuNetNAS: An Immune-network approach for searching Convolutional
Neural Network Architectures
- URL: http://arxiv.org/abs/2002.12704v1
- Date: Fri, 28 Feb 2020 13:32:57 GMT
- Title: ImmuNetNAS: An Immune-network approach for searching Convolutional
Neural Network Architectures
- Authors: Kefan Chen, Wei Pang
- Abstract summary: ImmuNetNAS is a novel Neural Architecture Search (NAS) approach inspired by the immune network theory.
The core of ImmuNetNAS is built on the original immune network algorithm, which iteratively updates the population through hypermutation and selection.
- Score: 9.880887106904517
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this research, we propose ImmuNetNAS, a novel Neural Architecture Search
(NAS) approach inspired by the immune network theory. The core of ImmuNetNAS is
built on the original immune network algorithm, which iteratively updates the
population through hypermutation and selection, and eliminates the
self-generation individuals that do not meet the requirements through comparing
antibody affinity and inter-specific similarity. In addition, in order to
facilitate the mutation operation, we propose a novel two-component based
neural structure coding strategy. Furthermore, an improved mutation strategy
based on Standard Genetic Algorithm (SGA) was proposed according to this
encoding method. Finally, based on the proposed two-component based coding
method, a new antibody affinity calculation method was developed to screen
suitable neural architectures. Systematic evaluations demonstrate that our
system has achieved good performance on both the MNIST and CIFAR-10 datasets.
We open-source our code on GitHub in order to share it with other deep learning
researchers and practitioners.
Related papers
- Stochastic Gradient Descent for Two-layer Neural Networks [2.0349026069285423]
This paper presents a study on the convergence rates of the descent (SGD) algorithm when applied to overparameterized two-layer neural networks.
Our approach combines the Tangent Kernel (NTK) approximation with convergence analysis in the Reproducing Kernel Space (RKHS) generated by NTK.
Our research framework enables us to explore the intricate interplay between kernel methods and optimization processes, shedding light on the dynamics and convergence properties of neural networks.
arXiv Detail & Related papers (2024-07-10T13:58:57Z) - Graph Neural Networks for Learning Equivariant Representations of Neural Networks [55.04145324152541]
We propose to represent neural networks as computational graphs of parameters.
Our approach enables a single model to encode neural computational graphs with diverse architectures.
We showcase the effectiveness of our method on a wide range of tasks, including classification and editing of implicit neural representations.
arXiv Detail & Related papers (2024-03-18T18:01:01Z) - Continuous Cartesian Genetic Programming based representation for
Multi-Objective Neural Architecture Search [12.545742558041583]
We propose a novel approach for designing less complex yet highly effective convolutional neural networks (CNNs)
Our approach combines real-based and block-chained CNNs representations based on cartesian genetic programming (CGP) for neural architecture search (NAS)
Two variants are introduced that differ in the granularity of the search space they consider.
arXiv Detail & Related papers (2023-06-05T07:32:47Z) - Permutation Equivariant Neural Functionals [92.0667671999604]
This work studies the design of neural networks that can process the weights or gradients of other neural networks.
We focus on the permutation symmetries that arise in the weights of deep feedforward networks because hidden layer neurons have no inherent order.
In our experiments, we find that permutation equivariant neural functionals are effective on a diverse set of tasks.
arXiv Detail & Related papers (2023-02-27T18:52:38Z) - Deep Convolutional Learning-Aided Detector for Generalized Frequency
Division Multiplexing with Index Modulation [0.0]
The proposed method first pre-processes the received signal by using a zero-forcing (ZF) detector and then uses a neural network consisting of a convolutional neural network (CNN) followed by a fully-connected neural network (FCNN)
The FCNN part uses only two fully-connected layers, which can be adapted to yield a trade-off between complexity and bit error rate (BER) performance.
It has been demonstrated that the proposed deep convolutional neural network-based detection and demodulation scheme provides better BER performance compared to ZF detector with a reasonable complexity increase.
arXiv Detail & Related papers (2022-02-06T22:18:42Z) - Detecting Communities in Complex Networks using an Adaptive Genetic
Algorithm and node similarity-based encoding [4.0714739042536845]
We propose a new node similarity-based encoding method to represent a network partition as an individual named MST-based.
Using the proposed method, we combine similarity-based and modularity-optimization-based approaches to find the communities of complex networks.
arXiv Detail & Related papers (2022-01-24T09:06:40Z) - Learning Autonomy in Management of Wireless Random Networks [102.02142856863563]
This paper presents a machine learning strategy that tackles a distributed optimization task in a wireless network with an arbitrary number of randomly interconnected nodes.
We develop a flexible deep neural network formalism termed distributed message-passing neural network (DMPNN) with forward and backward computations independent of the network topology.
arXiv Detail & Related papers (2021-06-15T09:03:28Z) - LocalDrop: A Hybrid Regularization for Deep Neural Networks [98.30782118441158]
We propose a new approach for the regularization of neural networks by the local Rademacher complexity called LocalDrop.
A new regularization function for both fully-connected networks (FCNs) and convolutional neural networks (CNNs) has been developed based on the proposed upper bound of the local Rademacher complexity.
arXiv Detail & Related papers (2021-03-01T03:10:11Z) - Graph Neural Network Encoding for Community Detection in Attribute
Networks [27.099775681534652]
A graph neural network encoding method is proposed to handle the community detection problem in complex attribute networks.
Experimental results show that the developed algorithm performs significantly better than some well-known evolutionary and non-evolutionary based algorithms.
arXiv Detail & Related papers (2020-06-06T23:41:36Z) - Binarizing MobileNet via Evolution-based Searching [66.94247681870125]
We propose a use of evolutionary search to facilitate the construction and training scheme when binarizing MobileNet.
Inspired by one-shot architecture search frameworks, we manipulate the idea of group convolution to design efficient 1-Bit Convolutional Neural Networks (CNNs)
Our objective is to come up with a tiny yet efficient binary neural architecture by exploring the best candidates of the group convolution.
arXiv Detail & Related papers (2020-05-13T13:25:51Z) - MSE-Optimal Neural Network Initialization via Layer Fusion [68.72356718879428]
Deep neural networks achieve state-of-the-art performance for a range of classification and inference tasks.
The use of gradient combined nonvolutionity renders learning susceptible to novel problems.
We propose fusing neighboring layers of deeper networks that are trained with random variables.
arXiv Detail & Related papers (2020-01-28T18:25:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.