Auto-Spikformer: Spikformer Architecture Search
- URL: http://arxiv.org/abs/2306.00807v1
- Date: Thu, 1 Jun 2023 15:35:26 GMT
- Title: Auto-Spikformer: Spikformer Architecture Search
- Authors: Kaiwei Che, Zhaokun Zhou, Zhengyu Ma, Wei Fang, Yanqi Chen, Shuaijie
Shen, Li Yuan, Yonghong Tian
- Abstract summary: Self-attention mechanisms have been integrated into Spiking Neural Networks (SNNs)
Recent advancements in SNN architecture, such as Spikformer, have demonstrated promising outcomes.
We propose Auto-Spikformer, a one-shot Transformer Architecture Search (TAS) method, which automates the quest for an optimized Spikformer architecture.
- Score: 22.332981906087785
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The integration of self-attention mechanisms into Spiking Neural Networks
(SNNs) has garnered considerable interest in the realm of advanced deep
learning, primarily due to their biological properties. Recent advancements in
SNN architecture, such as Spikformer, have demonstrated promising outcomes by
leveraging Spiking Self-Attention (SSA) and Spiking Patch Splitting (SPS)
modules. However, we observe that Spikformer may exhibit excessive energy
consumption, potentially attributable to redundant channels and blocks. To
mitigate this issue, we propose Auto-Spikformer, a one-shot Transformer
Architecture Search (TAS) method, which automates the quest for an optimized
Spikformer architecture. To facilitate the search process, we propose methods
Evolutionary SNN neurons (ESNN), which optimizes the SNN parameters, and apply
the previous method of weight entanglement supernet training, which optimizes
the Vision Transformer (ViT) parameters. Moreover, we propose an accuracy and
energy balanced fitness function $\mathcal{F}_{AEB}$ that jointly considers
both energy consumption and accuracy, and aims to find a Pareto optimal
combination that balances these two objectives. Our experimental results
demonstrate the effectiveness of Auto-Spikformer, which outperforms the
state-of-the-art method including CNN or ViT models that are manually or
automatically designed while significantly reducing energy consumption.
Related papers
- Q-SNNs: Quantized Spiking Neural Networks [12.719590949933105]
Spiking Neural Networks (SNNs) leverage sparse spikes to represent information and process them in an event-driven manner.
We introduce a lightweight and hardware-friendly Quantized SNN that applies quantization to both synaptic weights and membrane potentials.
We present a new Weight-Spike Dual Regulation (WS-DR) method inspired by information entropy theory.
arXiv Detail & Related papers (2024-06-19T16:23:26Z) - SpikingResformer: Bridging ResNet and Vision Transformer in Spiking Neural Networks [22.665939536001797]
We propose a novel spiking self-attention mechanism named Dual Spike Self-Attention (DSSA) with a reasonable scaling method.
Based on DSSA, we propose a novel spiking Vision Transformer architecture called SpikingResformer.
We show that SpikingResformer achieves higher accuracy with fewer parameters and lower energy consumption than other spiking Vision Transformer counterparts.
arXiv Detail & Related papers (2024-03-21T11:16:42Z) - Spiking-PhysFormer: Camera-Based Remote Photoplethysmography with Parallel Spike-driven Transformer [15.08113674331192]
Spiking networks (SNNs) hold immense potential for energy-efficient deep learning.
We propose a hybrid neural network (HNN) model, the Spiking-PhysFormer, aimed at reducing power consumption.
The proposed model achieves a 12.4% reduction in power consumption compared to PhysFormer.
arXiv Detail & Related papers (2024-02-07T12:38:47Z) - LitE-SNN: Designing Lightweight and Efficient Spiking Neural Network through Spatial-Temporal Compressive Network Search and Joint Optimization [48.41286573672824]
Spiking Neural Networks (SNNs) mimic the information-processing mechanisms of the human brain and are highly energy-efficient.
We propose a new approach named LitE-SNN that incorporates both spatial and temporal compression into the automated network design process.
arXiv Detail & Related papers (2024-01-26T05:23:11Z) - SpikingJelly: An open-source machine learning infrastructure platform
for spike-based intelligence [51.6943465041708]
Spiking neural networks (SNNs) aim to realize brain-inspired intelligence on neuromorphic chips with high energy efficiency.
We contribute a full-stack toolkit for pre-processing neuromorphic datasets, building deep SNNs, optimizing their parameters, and deploying SNNs on neuromorphic chips.
arXiv Detail & Related papers (2023-10-25T13:15:17Z) - AutoST: Training-free Neural Architecture Search for Spiking
Transformers [14.791412391584064]
Spiking Transformers achieve both the energy efficiency of Spiking Neural Networks (SNNs) and the high capacity of Transformers.
Existing Spiking Transformer architectures exhibit a notable architectural gap, resulting in suboptimal performance.
We introduce AutoST, a training-free NAS method for Spiking Transformers, to rapidly identify high-performance Spiking Transformer architectures.
arXiv Detail & Related papers (2023-07-01T10:19:52Z) - Systematic Architectural Design of Scale Transformed Attention Condenser
DNNs via Multi-Scale Class Representational Response Similarity Analysis [93.0013343535411]
We propose a novel type of analysis called Multi-Scale Class Representational Response Similarity Analysis (ClassRepSim)
We show that adding STAC modules to ResNet style architectures can result in up to a 1.6% increase in top-1 accuracy.
Results from ClassRepSim analysis can be used to select an effective parameterization of the STAC module resulting in competitive performance.
arXiv Detail & Related papers (2023-06-16T18:29:26Z) - A Generic Shared Attention Mechanism for Various Backbone Neural Networks [53.36677373145012]
Self-attention modules (SAMs) produce strongly correlated attention maps across different layers.
Dense-and-Implicit Attention (DIA) shares SAMs across layers and employs a long short-term memory module.
Our simple yet effective DIA can consistently enhance various network backbones.
arXiv Detail & Related papers (2022-10-27T13:24:08Z) - Spikformer: When Spiking Neural Network Meets Transformer [102.91330530210037]
We consider two biologically plausible structures, the Spiking Neural Network (SNN) and the self-attention mechanism.
We propose a novel Spiking Self Attention (SSA) as well as a powerful framework, named Spiking Transformer (Spikformer)
arXiv Detail & Related papers (2022-09-29T14:16:49Z) - Auto-PINN: Understanding and Optimizing Physics-Informed Neural
Architecture [77.59766598165551]
Physics-informed neural networks (PINNs) are revolutionizing science and engineering practice by bringing together the power of deep learning to bear on scientific computation.
Here, we propose Auto-PINN, which employs Neural Architecture Search (NAS) techniques to PINN design.
A comprehensive set of pre-experiments using standard PDE benchmarks allows us to probe the structure-performance relationship in PINNs.
arXiv Detail & Related papers (2022-05-27T03:24:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.