Examining the causal structures of deep neural networks using
information theory
- URL: http://arxiv.org/abs/2010.13871v1
- Date: Mon, 26 Oct 2020 19:53:16 GMT
- Title: Examining the causal structures of deep neural networks using
information theory
- Authors: Simon Mattsson, Eric J. Michaud, Erik Hoel
- Abstract summary: Deep Neural Networks (DNNs) are often examined at the level of their response to input, such as analyzing the mutual information between nodes and data sets.
DNNs can also be examined at the level of causation, exploring "what does what" within the layers of the network itself.
Here, we introduce a suite of metrics based on information theory to quantify and track changes in the causal structure of DNNs during training.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep Neural Networks (DNNs) are often examined at the level of their response
to input, such as analyzing the mutual information between nodes and data sets.
Yet DNNs can also be examined at the level of causation, exploring "what does
what" within the layers of the network itself. Historically, analyzing the
causal structure of DNNs has received less attention than understanding their
responses to input. Yet definitionally, generalizability must be a function of
a DNN's causal structure since it reflects how the DNN responds to unseen or
even not-yet-defined future inputs. Here, we introduce a suite of metrics based
on information theory to quantify and track changes in the causal structure of
DNNs during training. Specifically, we introduce the effective information (EI)
of a feedforward DNN, which is the mutual information between layer input and
output following a maximum-entropy perturbation. The EI can be used to assess
the degree of causal influence nodes and edges have over their downstream
targets in each layer. We show that the EI can be further decomposed in order
to examine the sensitivity of a layer (measured by how well edges transmit
perturbations) and the degeneracy of a layer (measured by how edge overlap
interferes with transmission), along with estimates of the amount of integrated
information of a layer. Together, these properties define where each layer lies
in the "causal plane" which can be used to visualize how layer connectivity
becomes more sensitive or degenerate over time, and how integration changes
during training, revealing how the layer-by-layer causal structure
differentiates. These results may help in understanding the generalization
capabilities of DNNs and provide foundational tools for making DNNs both more
generalizable and more explainable.
Related papers
- Neural Networks Decoded: Targeted and Robust Analysis of Neural Network Decisions via Causal Explanations and Reasoning [9.947555560412397]
We introduce TRACER, a novel method grounded in causal inference theory to estimate the causal dynamics underpinning DNN decisions.
Our approach systematically intervenes on input features to observe how specific changes propagate through the network, affecting internal activations and final outputs.
TRACER further enhances explainability by generating counterfactuals that reveal possible model biases and offer contrastive explanations for misclassifications.
arXiv Detail & Related papers (2024-10-07T20:44:53Z) - Information-Theoretic Generalization Bounds for Deep Neural Networks [22.87479366196215]
Deep neural networks (DNNs) exhibit an exceptional capacity for generalization in practical applications.
This work aims to capture the effect and benefits of depth for supervised learning via information-theoretic generalization bounds.
arXiv Detail & Related papers (2024-04-04T03:20:35Z) - SAfER: Layer-Level Sensitivity Assessment for Efficient and Robust
Neural Network Inference [20.564198591600647]
Deep neural networks (DNNs) demonstrate outstanding performance across most computer vision tasks.
Some critical applications, such as autonomous driving or medical imaging, also require investigation into their behavior.
DNN attribution consists in studying the relationship between the predictions of a DNN and its inputs.
arXiv Detail & Related papers (2023-08-09T07:45:51Z) - Transferability of coVariance Neural Networks and Application to
Interpretable Brain Age Prediction using Anatomical Features [119.45320143101381]
Graph convolutional networks (GCN) leverage topology-driven graph convolutional operations to combine information across the graph for inference tasks.
We have studied GCNs with covariance matrices as graphs in the form of coVariance neural networks (VNNs)
VNNs inherit the scale-free data processing architecture from GCNs and here, we show that VNNs exhibit transferability of performance over datasets whose covariance matrices converge to a limit object.
arXiv Detail & Related papers (2023-05-02T22:15:54Z) - On the Intrinsic Structures of Spiking Neural Networks [66.57589494713515]
Recent years have emerged a surge of interest in SNNs owing to their remarkable potential to handle time-dependent and event-driven data.
There has been a dearth of comprehensive studies examining the impact of intrinsic structures within spiking computations.
This work delves deep into the intrinsic structures of SNNs, by elucidating their influence on the expressivity of SNNs.
arXiv Detail & Related papers (2022-06-21T09:42:30Z) - Generalization ability and Vulnerabilities to adversarial perturbations: Two sides of the same coin [1.7314342339585087]
We use the self-organizing map (SOM) to analyze internal codes associated with deep learning models' decision-making.
Our analyses suggest that shallow layers close to the input layer map onto homogeneous codes and that deep layers close to the output layer transform these homogeneous codes in shallow layers to diverse codes.
arXiv Detail & Related papers (2022-05-22T23:14:27Z) - Deep Architecture Connectivity Matters for Its Convergence: A
Fine-Grained Analysis [94.64007376939735]
We theoretically characterize the impact of connectivity patterns on the convergence of deep neural networks (DNNs) under gradient descent training.
We show that by a simple filtration on "unpromising" connectivity patterns, we can trim down the number of models to evaluate.
arXiv Detail & Related papers (2022-05-11T17:43:54Z) - Understanding and Diagnosing Vulnerability under Adversarial Attacks [62.661498155101654]
Deep Neural Networks (DNNs) are known to be vulnerable to adversarial attacks.
We propose a novel interpretability method, InterpretGAN, to generate explanations for features used for classification in latent variables.
We also design the first diagnostic method to quantify the vulnerability contributed by each layer.
arXiv Detail & Related papers (2020-07-17T01:56:28Z) - Boosting Deep Neural Networks with Geometrical Prior Knowledge: A Survey [77.99182201815763]
Deep Neural Networks (DNNs) achieve state-of-the-art results in many different problem settings.
DNNs are often treated as black box systems, which complicates their evaluation and validation.
One promising field, inspired by the success of convolutional neural networks (CNNs) in computer vision tasks, is to incorporate knowledge about symmetric geometrical transformations.
arXiv Detail & Related papers (2020-06-30T14:56:05Z) - Architecture Disentanglement for Deep Neural Networks [174.16176919145377]
We introduce neural architecture disentanglement (NAD) to explain the inner workings of deep neural networks (DNNs)
NAD learns to disentangle a pre-trained DNN into sub-architectures according to independent tasks, forming information flows that describe the inference processes.
Results show that misclassified images have a high probability of being assigned to task sub-architectures similar to the correct ones.
arXiv Detail & Related papers (2020-03-30T08:34:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.