MILR: Mathematically Induced Layer Recovery for Plaintext Space Error
Correction of CNNs
- URL: http://arxiv.org/abs/2010.14687v1
- Date: Wed, 28 Oct 2020 00:47:15 GMT
- Title: MILR: Mathematically Induced Layer Recovery for Plaintext Space Error
Correction of CNNs
- Authors: Jonathan Ponader, Sandip Kundu, Yan Solihin
- Abstract summary: This paper proposes MILR, a software based CNN error detection and error correction system.
The self-healing capabilities are based on mathematical relationships between the inputs,outputs, and parameters(weights) of a layers.
- Score: 4.23546023847456
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The increased use of Convolutional Neural Networks (CNN) in mission critical
systems has increased the need for robust and resilient networks in the face of
both naturally occurring faults as well as security attacks. The lack of
robustness and resiliency can lead to unreliable inference results. Current
methods that address CNN robustness require hardware modification, network
modification, or network duplication. This paper proposes MILR a software based
CNN error detection and error correction system that enables self-healing of
the network from single and multi bit errors. The self-healing capabilities are
based on mathematical relationships between the inputs,outputs, and
parameters(weights) of a layers, exploiting these relationships allow the
recovery of erroneous parameters (weights) throughout a layer and the network.
MILR is suitable for plaintext-space error correction (PSEC) given its ability
to correct whole-weight and even whole-layer errors in CNNs.
Related papers
- Physical Rule-Guided Convolutional Neural Network [0.0]
Physics-Guided Neural Networks (PGNNs) have emerged to address limitations by integrating scientific principles and real-world knowledge.
This paper proposes a novel Physics-Guided CNN (PGCNN) architecture that incorporates dynamic, trainable, and automated LLM-generated, widely recognized rules integrated into the model as custom layers.
The PGCNN is evaluated on multiple datasets, demonstrating superior performance compared to a baseline CNN model.
arXiv Detail & Related papers (2024-09-03T17:32:35Z) - Cost-Effective Fault Tolerance for CNNs Using Parameter Vulnerability Based Hardening and Pruning [0.4660328753262075]
This paper introduces a model-level hardening approach for CNNs by integrating error correction directly into the neural networks.
The proposed method demonstrates fault resilience nearly equivalent to TMR-based correction but with significantly reduced overhead.
Remarkably, the hardened pruned CNNs perform up to 24% faster than the hardened un-pruned ones.
arXiv Detail & Related papers (2024-05-17T09:42:44Z) - Dynamic Semantic Compression for CNN Inference in Multi-access Edge
Computing: A Graph Reinforcement Learning-based Autoencoder [82.8833476520429]
We propose a novel semantic compression method, autoencoder-based CNN architecture (AECNN) for effective semantic extraction and compression in partial offloading.
In the semantic encoder, we introduce a feature compression module based on the channel attention mechanism in CNNs, to compress intermediate data by selecting the most informative features.
In the semantic decoder, we design a lightweight decoder to reconstruct the intermediate data through learning from the received compressed data to improve accuracy.
arXiv Detail & Related papers (2024-01-19T15:19:47Z) - Certified Invertibility in Neural Networks via Mixed-Integer Programming [16.64960701212292]
Neural networks are known to be vulnerable to adversarial attacks.
There may exist large, meaningful perturbations that do not affect the network's decision.
We discuss how our findings can be useful for invertibility certification in transformations between neural networks.
arXiv Detail & Related papers (2023-01-27T15:40:38Z) - Fast and Accurate Error Simulation for CNNs against Soft Errors [64.54260986994163]
We present a framework for the reliability analysis of Conal Neural Networks (CNNs) via an error simulation engine.
These error models are defined based on the corruption patterns of the output of the CNN operators induced by faults.
We show that our methodology achieves about 99% accuracy of the fault effects w.r.t. SASSIFI, and a speedup ranging from 44x up to 63x w.r.t.FI, that only implements a limited set of error models.
arXiv Detail & Related papers (2022-06-04T19:45:02Z) - Performance and accuracy assessments of an incompressible fluid solver
coupled with a deep Convolutional Neural Network [0.0]
The resolution of the Poisson equation is usually one of the most computationally intensive steps for incompressible fluid solvers.
CNN has been introduced to solve this equation, leading to significant inference time reduction.
A hybrid strategy is developed, which couples a CNN with a traditional iterative solver to ensure a user-defined accuracy level.
arXiv Detail & Related papers (2021-09-20T08:30:29Z) - Neural Architecture Dilation for Adversarial Robustness [56.18555072877193]
A shortcoming of convolutional neural networks is that they are vulnerable to adversarial attacks.
This paper aims to improve the adversarial robustness of the backbone CNNs that have a satisfactory accuracy.
Under a minimal computational overhead, a dilation architecture is expected to be friendly with the standard performance of the backbone CNN.
arXiv Detail & Related papers (2021-08-16T03:58:00Z) - Manipulating Identical Filter Redundancy for Efficient Pruning on Deep
and Complicated CNN [126.88224745942456]
We propose a novel Centripetal SGD (C-SGD) to make some filters identical, resulting in ideal redundancy patterns.
C-SGD delivers better performance because the redundancy is better organized, compared to the existing methods.
arXiv Detail & Related papers (2021-07-30T06:18:19Z) - BreakingBED -- Breaking Binary and Efficient Deep Neural Networks by
Adversarial Attacks [65.2021953284622]
We study robustness of CNNs against white-box and black-box adversarial attacks.
Results are shown for distilled CNNs, agent-based state-of-the-art pruned models, and binarized neural networks.
arXiv Detail & Related papers (2021-03-14T20:43:19Z) - FT-CNN: Algorithm-Based Fault Tolerance for Convolutional Neural
Networks [13.100954947774163]
Convolutional neural networks (CNNs) are becoming more and more important for solving challenging and critical problems in many fields.
CNN inference applications have been deployed in safety-critical systems, which may suffer from soft errors caused by high-energy particles, high temperature, or abnormal voltage.
Traditional fault tolerance methods are not suitable for CNN inference because error-correcting code is unable to protect computational components.
arXiv Detail & Related papers (2020-03-27T02:01:54Z) - Approximation and Non-parametric Estimation of ResNet-type Convolutional
Neural Networks [52.972605601174955]
We show a ResNet-type CNN can attain the minimax optimal error rates in important function classes.
We derive approximation and estimation error rates of the aformentioned type of CNNs for the Barron and H"older classes.
arXiv Detail & Related papers (2019-03-24T19:42:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.