A Neural Architecture Search based Framework for Liquid State Machine
Design
- URL: http://arxiv.org/abs/2004.07864v1
- Date: Tue, 7 Apr 2020 10:55:05 GMT
- Title: A Neural Architecture Search based Framework for Liquid State Machine
Design
- Authors: Shuo Tian, Lianhua Qu, Kai Hu, Nan Li, Lei Wang and Weixia Xu
- Abstract summary: Liquid State Machine (LSM) is a recurrent version of Spiking Neural Networks (SNN)
Recent works have demonstrated great potential for improving the accuracy of LSM model with low complexity.
Considering the diversity and uniqueness of brain structure, the design of LSM model should be explored in the largest search space possible.
- Score: 7.729541832738546
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Liquid State Machine (LSM), also known as the recurrent version of Spiking
Neural Networks (SNN), has attracted great research interests thanks to its
high computational power, biological plausibility from the brain, simple
structure and low training complexity. By exploring the design space in network
architectures and parameters, recent works have demonstrated great potential
for improving the accuracy of LSM model with low complexity. However, these
works are based on manually-defined network architectures or predefined
parameters. Considering the diversity and uniqueness of brain structure, the
design of LSM model should be explored in the largest search space possible. In
this paper, we propose a Neural Architecture Search (NAS) based framework to
explore both architecture and parameter design space for automatic
dataset-oriented LSM model. To handle the exponentially-increased design space,
we adopt a three-step search for LSM, including multi-liquid architecture
search, variation on the number of neurons and parameters search such as
percentage connectivity and excitatory neuron ratio within each liquid.
Besides, we propose to use Simulated Annealing (SA) algorithm to implement the
three-step heuristic search. Three datasets, including image dataset of MNIST
and NMNIST and speech dataset of FSDD, are used to test the effectiveness of
our proposed framework. Simulation results show that our proposed framework can
produce the dataset-oriented optimal LSM models with high accuracy and low
complexity. The best classification accuracy on the three datasets is 93.2%,
92.5% and 84% respectively with only 1000 spiking neurons, and the network
connections can be averagely reduced by 61.4% compared with a single LSM.
Moreover, we find that the total quantity of neurons in optimal LSM models on
three datasets can be further reduced by 20% with only about 0.5% accuracy
loss.
Related papers
- Differentiable architecture search with multi-dimensional attention for spiking neural networks [4.318876451929319]
Spiking Neural Networks (SNNs) have gained enormous popularity in the field of artificial intelligence.
The majority of SNN methods directly inherit the structure of Artificial Neural Networks (ANN)
We propose Multi-Attention Differentiable Architecture Search (MA-DARTS) to directly automate the search for the optimal network structure of SNNs.
arXiv Detail & Related papers (2024-11-01T07:18:32Z) - A Pairwise Comparison Relation-assisted Multi-objective Evolutionary Neural Architecture Search Method with Multi-population Mechanism [58.855741970337675]
Neural architecture search (NAS) enables re-searchers to automatically explore vast search spaces and find efficient neural networks.
NAS suffers from a key bottleneck, i.e., numerous architectures need to be evaluated during the search process.
We propose the SMEM-NAS, a pairwise com-parison relation-assisted multi-objective evolutionary algorithm based on a multi-population mechanism.
arXiv Detail & Related papers (2024-07-22T12:46:22Z) - Multi-Objective Neural Architecture Search for In-Memory Computing [0.5892638927736115]
We employ neural architecture search (NAS) to enhance the efficiency of deploying diverse machine learning (ML) tasks on in-memory computing architectures.
Our evaluation of this NAS approach for IMC architecture deployment spans three distinct image classification datasets.
arXiv Detail & Related papers (2024-06-10T19:17:09Z) - Multi-scale Evolutionary Neural Architecture Search for Deep Spiking
Neural Networks [7.271032282434803]
We propose a Multi-Scale Evolutionary Neural Architecture Search (MSE-NAS) for Spiking Neural Networks (SNNs)
MSE-NAS evolves individual neuron operation, self-organized integration of multiple circuit motifs, and global connectivity across motifs through a brain-inspired indirect evaluation function, Representational Dissimilarity Matrices (RDMs)
The proposed algorithm achieves state-of-the-art (SOTA) performance with shorter simulation steps on static datasets and neuromorphic datasets.
arXiv Detail & Related papers (2023-04-21T05:36:37Z) - A Local Optima Network Analysis of the Feedforward Neural Architecture
Space [0.0]
Local optima network (LON) analysis is a derivative of the fitness landscape of candidate solutions.
LONs may provide a viable paradigm by which to analyse and optimise neural architectures.
arXiv Detail & Related papers (2022-06-02T08:09:17Z) - ANNETTE: Accurate Neural Network Execution Time Estimation with Stacked
Models [56.21470608621633]
We propose a time estimation framework to decouple the architectural search from the target hardware.
The proposed methodology extracts a set of models from micro- kernel and multi-layer benchmarks and generates a stacked model for mapping and network execution time estimation.
We compare estimation accuracy and fidelity of the generated mixed models, statistical models with the roofline model, and a refined roofline model for evaluation.
arXiv Detail & Related papers (2021-05-07T11:39:05Z) - Differentiable Neural Architecture Learning for Efficient Neural Network
Design [31.23038136038325]
We introduce a novel emph architecture parameterisation based on scaled sigmoid function.
We then propose a general emphiable Neural Architecture Learning (DNAL) method to optimize the neural architecture without the need to evaluate candidate neural networks.
arXiv Detail & Related papers (2021-03-03T02:03:08Z) - MS-RANAS: Multi-Scale Resource-Aware Neural Architecture Search [94.80212602202518]
We propose Multi-Scale Resource-Aware Neural Architecture Search (MS-RANAS)
We employ a one-shot architecture search approach in order to obtain a reduced search cost.
We achieve state-of-the-art results in terms of accuracy-speed trade-off.
arXiv Detail & Related papers (2020-09-29T11:56:01Z) - DrNAS: Dirichlet Neural Architecture Search [88.56953713817545]
We treat the continuously relaxed architecture mixing weight as random variables, modeled by Dirichlet distribution.
With recently developed pathwise derivatives, the Dirichlet parameters can be easily optimized with gradient-based generalization.
To alleviate the large memory consumption of differentiable NAS, we propose a simple yet effective progressive learning scheme.
arXiv Detail & Related papers (2020-06-18T08:23:02Z) - FBNetV3: Joint Architecture-Recipe Search using Predictor Pretraining [65.39532971991778]
We present an accuracy predictor that scores architecture and training recipes jointly, guiding both sample selection and ranking.
We run fast evolutionary searches in just CPU minutes to generate architecture-recipe pairs for a variety of resource constraints.
FBNetV3 makes up a family of state-of-the-art compact neural networks that outperform both automatically and manually-designed competitors.
arXiv Detail & Related papers (2020-06-03T05:20:21Z) - When Residual Learning Meets Dense Aggregation: Rethinking the
Aggregation of Deep Neural Networks [57.0502745301132]
We propose Micro-Dense Nets, a novel architecture with global residual learning and local micro-dense aggregations.
Our micro-dense block can be integrated with neural architecture search based models to boost their performance.
arXiv Detail & Related papers (2020-04-19T08:34:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.