Analysis of Power-Oriented Fault Injection Attacks on Spiking Neural
Networks
- URL: http://arxiv.org/abs/2204.04768v1
- Date: Sun, 10 Apr 2022 20:48:46 GMT
- Title: Analysis of Power-Oriented Fault Injection Attacks on Spiking Neural
Networks
- Authors: Karthikeyan Nagarajan, Junde Li, Sina Sayyah Ensan, Mohammad Nasim
Imtiaz Khan, Sachhidh Kannan, and Swaroop Ghosh
- Abstract summary: Spiking Neural Networks (SNNs) are quickly gaining traction as a viable alternative to Deep Neural Networks (DNNs)
SNNs contain security-sensitive assets (e.g., neuron threshold voltage) and vulnerabilities that adversaries can exploit.
We investigate global fault injection attacks by employing external power supplies and laser-induced local power glitches.
We find that in the worst-case scenario, classification accuracy is reduced by 85.65%.
- Score: 5.7494562086770955
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Spiking Neural Networks (SNN) are quickly gaining traction as a viable
alternative to Deep Neural Networks (DNN). In comparison to DNNs, SNNs are more
computationally powerful and provide superior energy efficiency. SNNs, while
exciting at first appearance, contain security-sensitive assets (e.g., neuron
threshold voltage) and vulnerabilities (e.g., sensitivity of classification
accuracy to neuron threshold voltage change) that adversaries can exploit. We
investigate global fault injection attacks by employing external power supplies
and laser-induced local power glitches to corrupt crucial training parameters
such as spike amplitude and neuron's membrane threshold potential on SNNs
developed using common analog neurons. We also evaluate the impact of
power-based attacks on individual SNN layers for 0% (i.e., no attack) to 100%
(i.e., whole layer under attack). We investigate the impact of the attacks on
digit classification tasks and find that in the worst-case scenario,
classification accuracy is reduced by 85.65%. We also propose defenses e.g., a
robust current driver design that is immune to power-oriented attacks, improved
circuit sizing of neuron components to reduce/recover the adversarial accuracy
degradation at the cost of negligible area and 25% power overhead. We also
present a dummy neuron-based voltage fault injection detection system with 1%
power and area overhead.
Related papers
- Fully Spiking Actor Network with Intra-layer Connections for
Reinforcement Learning [51.386945803485084]
We focus on the task where the agent needs to learn multi-dimensional deterministic policies to control.
Most existing spike-based RL methods take the firing rate as the output of SNNs, and convert it to represent continuous action space (i.e., the deterministic policy) through a fully-connected layer.
To develop a fully spiking actor network without any floating-point matrix operations, we draw inspiration from the non-spiking interneurons found in insects.
arXiv Detail & Related papers (2024-01-09T07:31:34Z) - Sneaky Spikes: Uncovering Stealthy Backdoor Attacks in Spiking Neural
Networks with Neuromorphic Data [15.084703823643311]
spiking neural networks (SNNs) offer enhanced energy efficiency and biologically plausible data processing capabilities.
This paper delves into backdoor attacks in SNNs using neuromorphic datasets and diverse triggers.
We present various attack strategies, achieving an attack success rate of up to 100% while maintaining a negligible impact on clean accuracy.
arXiv Detail & Related papers (2023-02-13T11:34:17Z) - Adversarial Defense via Neural Oscillation inspired Gradient Masking [0.0]
Spiking neural networks (SNNs) attract great attention due to their low power consumption, low latency, and biological plausibility.
We propose a novel neural model that incorporates the bio-inspired oscillation mechanism to enhance the security of SNNs.
arXiv Detail & Related papers (2022-11-04T02:13:19Z) - Training High-Performance Low-Latency Spiking Neural Networks by
Differentiation on Spike Representation [70.75043144299168]
Spiking Neural Network (SNN) is a promising energy-efficient AI model when implemented on neuromorphic hardware.
It is a challenge to efficiently train SNNs due to their non-differentiability.
We propose the Differentiation on Spike Representation (DSR) method, which could achieve high performance.
arXiv Detail & Related papers (2022-05-01T12:44:49Z) - Improving Adversarial Transferability via Neuron Attribution-Based
Attacks [35.02147088207232]
We propose the Neuron-based Attack (NAA), which conducts feature-level attacks with more accurate neuron importance estimations.
We derive an approximation scheme of neuron attribution to tremendously reduce the overhead.
Experiments confirm the superiority of our approach to the state-of-the-art benchmarks.
arXiv Detail & Related papers (2022-03-31T13:47:30Z) - Energy-Efficient High-Accuracy Spiking Neural Network Inference Using
Time-Domain Neurons [0.18352113484137625]
This paper presents a low-power highly linear time-domain I&F neuron circuit.
The proposed neuron leads to more than 4.3x lower error rate on the MNIST inference.
The power consumed by the proposed neuron circuit is simulated to be 0.230uW per neuron, which is orders of magnitude lower than the existing voltage-domain neurons.
arXiv Detail & Related papers (2022-02-04T08:24:03Z) - Neural Architecture Dilation for Adversarial Robustness [56.18555072877193]
A shortcoming of convolutional neural networks is that they are vulnerable to adversarial attacks.
This paper aims to improve the adversarial robustness of the backbone CNNs that have a satisfactory accuracy.
Under a minimal computational overhead, a dilation architecture is expected to be friendly with the standard performance of the backbone CNN.
arXiv Detail & Related papers (2021-08-16T03:58:00Z) - BreakingBED -- Breaking Binary and Efficient Deep Neural Networks by
Adversarial Attacks [65.2021953284622]
We study robustness of CNNs against white-box and black-box adversarial attacks.
Results are shown for distilled CNNs, agent-based state-of-the-art pruned models, and binarized neural networks.
arXiv Detail & Related papers (2021-03-14T20:43:19Z) - And/or trade-off in artificial neurons: impact on adversarial robustness [91.3755431537592]
Presence of sufficient number of OR-like neurons in a network can lead to classification brittleness and increased vulnerability to adversarial attacks.
We define AND-like neurons and propose measures to increase their proportion in the network.
Experimental results on the MNIST dataset suggest that our approach holds promise as a direction for further exploration.
arXiv Detail & Related papers (2021-02-15T08:19:05Z) - Inherent Adversarial Robustness of Deep Spiking Neural Networks: Effects
of Discrete Input Encoding and Non-Linear Activations [9.092733355328251]
Spiking Neural Network (SNN) is a potential candidate for inherent robustness against adversarial attacks.
In this work, we demonstrate that adversarial accuracy of SNNs under gradient-based attacks is higher than their non-spiking counterparts.
arXiv Detail & Related papers (2020-03-23T17:20:24Z) - Non-linear Neurons with Human-like Apical Dendrite Activations [81.18416067005538]
We show that a standard neuron followed by our novel apical dendrite activation (ADA) can learn the XOR logical function with 100% accuracy.
We conduct experiments on six benchmark data sets from computer vision, signal processing and natural language processing.
arXiv Detail & Related papers (2020-02-02T21:09:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.