Explainability Tools Enabling Deep Learning in Future In-Situ Real-Time
Planetary Explorations
- URL: http://arxiv.org/abs/2201.05775v1
- Date: Sat, 15 Jan 2022 07:10:00 GMT
- Title: Explainability Tools Enabling Deep Learning in Future In-Situ Real-Time
Planetary Explorations
- Authors: Daniel Lundstrom, Alexander Huyen, Arya Mevada, Kyongsik Yun, Thomas
Lu
- Abstract summary: Deep learning (DL) has proven to be an effective machine learning and computer vision technique.
Most of the Deep Neural Network (DNN) architectures are so complex that they are considered a 'black box'
In this paper, we used integrated gradients to describe the attributions of each neuron to the output classes.
It provides a set of explainability tools (ET) that opens the black box of a DNN so that the individual contribution of neurons to category classification can be ranked and visualized.
- Score: 58.720142291102135
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep learning (DL) has proven to be an effective machine learning and
computer vision technique. DL-based image segmentation, object recognition and
classification will aid many in-situ Mars rover tasks such as path planning and
artifact recognition/extraction. However, most of the Deep Neural Network (DNN)
architectures are so complex that they are considered a 'black box'. In this
paper, we used integrated gradients to describe the attributions of each neuron
to the output classes. It provides a set of explainability tools (ET) that
opens the black box of a DNN so that the individual contribution of neurons to
category classification can be ranked and visualized. The neurons in each dense
layer are mapped and ranked by measuring expected contribution of a neuron to a
class vote given a true image label. The importance of neurons is prioritized
according to their correct or incorrect contribution to the output classes and
suppression or bolstering of incorrect classes, weighted by the size of each
class. ET provides an interface to prune the network to enhance high-rank
neurons and remove low-performing neurons. ET technology will make DNNs smaller
and more efficient for implementation in small embedded systems. It also leads
to more explainable and testable DNNs that can make systems easier for
Validation \& Verification. The goal of ET technology is to enable the adoption
of DL in future in-situ planetary exploration missions.
Related papers
- Fully Spiking Actor Network with Intra-layer Connections for
Reinforcement Learning [51.386945803485084]
We focus on the task where the agent needs to learn multi-dimensional deterministic policies to control.
Most existing spike-based RL methods take the firing rate as the output of SNNs, and convert it to represent continuous action space (i.e., the deterministic policy) through a fully-connected layer.
To develop a fully spiking actor network without any floating-point matrix operations, we draw inspiration from the non-spiking interneurons found in insects.
arXiv Detail & Related papers (2024-01-09T07:31:34Z) - Curriculum Design Helps Spiking Neural Networks to Classify Time Series [16.402675046686834]
Spiking Neural Networks (SNNs) have a greater potential for modeling time series data than Artificial Neural Networks (ANNs)
In this work, enlighten by brain-inspired science, we find that, not only the structure but also the learning process should be human-like.
arXiv Detail & Related papers (2023-12-26T02:04:53Z) - A Hybrid Neural Coding Approach for Pattern Recognition with Spiking
Neural Networks [53.31941519245432]
Brain-inspired spiking neural networks (SNNs) have demonstrated promising capabilities in solving pattern recognition tasks.
These SNNs are grounded on homogeneous neurons that utilize a uniform neural coding for information representation.
In this study, we argue that SNN architectures should be holistically designed to incorporate heterogeneous coding schemes.
arXiv Detail & Related papers (2023-05-26T02:52:12Z) - Transferability of coVariance Neural Networks and Application to
Interpretable Brain Age Prediction using Anatomical Features [119.45320143101381]
Graph convolutional networks (GCN) leverage topology-driven graph convolutional operations to combine information across the graph for inference tasks.
We have studied GCNs with covariance matrices as graphs in the form of coVariance neural networks (VNNs)
VNNs inherit the scale-free data processing architecture from GCNs and here, we show that VNNs exhibit transferability of performance over datasets whose covariance matrices converge to a limit object.
arXiv Detail & Related papers (2023-05-02T22:15:54Z) - Symbiosis of an artificial neural network and models of biological
neurons: training and testing [0.0]
We show the possibility of creating and identifying the features of an artificial neural network (ANN) which consists of mathematical models of biological neurons.
The FitzHugh--Nagumo (FHN) system is used as an example of model demonstrating simplified neuron activity.
arXiv Detail & Related papers (2023-02-03T10:06:54Z) - A Spiking Neural Network Structure Implementing Reinforcement Learning [0.0]
In the present paper, I describe an SNN structure which, seemingly, can be used in wide range of reinforcement learning tasks.
The SNN structure considered in the paper includes spiking neurons described by a generalization of the LIFAT (leaky integrate-and-fire neuron with adaptive threshold) model.
My concept is based on very general assumptions about RL task characteristics and has no visible limitations on its applicability.
arXiv Detail & Related papers (2022-04-09T09:08:10Z) - Visualizing Deep Neural Networks with Topographic Activation Maps [1.1470070927586014]
We introduce and compare methods to obtain a topographic layout of neurons in a Deep Neural Network layer.
We demonstrate how to use topographic activation maps to identify errors or encoded biases and to visualize training processes.
arXiv Detail & Related papers (2022-04-07T15:56:44Z) - Locality Guided Neural Networks for Explainable Artificial Intelligence [12.435539489388708]
We propose a novel algorithm for back propagation, called Locality Guided Neural Network(LGNN)
LGNN preserves locality between neighbouring neurons within each layer of a deep network.
In our experiments, we train various VGG and Wide ResNet (WRN) networks for image classification on CIFAR100.
arXiv Detail & Related papers (2020-07-12T23:45:51Z) - Progressive Tandem Learning for Pattern Recognition with Deep Spiking
Neural Networks [80.15411508088522]
Spiking neural networks (SNNs) have shown advantages over traditional artificial neural networks (ANNs) for low latency and high computational efficiency.
We propose a novel ANN-to-SNN conversion and layer-wise learning framework for rapid and efficient pattern recognition.
arXiv Detail & Related papers (2020-07-02T15:38:44Z) - Non-linear Neurons with Human-like Apical Dendrite Activations [81.18416067005538]
We show that a standard neuron followed by our novel apical dendrite activation (ADA) can learn the XOR logical function with 100% accuracy.
We conduct experiments on six benchmark data sets from computer vision, signal processing and natural language processing.
arXiv Detail & Related papers (2020-02-02T21:09:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.