Higher-order mutual information reveals synergistic sub-networks for
multi-neuron importance
- URL: http://arxiv.org/abs/2211.00416v1
- Date: Tue, 1 Nov 2022 12:21:15 GMT
- Title: Higher-order mutual information reveals synergistic sub-networks for
multi-neuron importance
- Authors: Kenzo Clauw, Sebastiano Stramaglia, Daniele Marinazzo
- Abstract summary: Previous work primarily attributed importance to individual neurons.
In this work, we study which groups of neurons contain synergistic or redundant information.
Results suggest our method can be used for pruning and unsupervised representation learning.
- Score: 0.483420384410068
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Quantifying which neurons are important with respect to the classification
decision of a trained neural network is essential for understanding their inner
workings. Previous work primarily attributed importance to individual neurons.
In this work, we study which groups of neurons contain synergistic or redundant
information using a multivariate mutual information method called the
O-information. We observe the first layer is dominated by redundancy suggesting
general shared features (i.e. detecting edges) while the last layer is
dominated by synergy indicating local class-specific features (i.e. concepts).
Finally, we show the O-information can be used for multi-neuron importance.
This can be demonstrated by re-training a synergistic sub-network, which
results in a minimal change in performance. These results suggest our method
can be used for pruning and unsupervised representation learning.
Related papers
- Towards Utilising a Range of Neural Activations for Comprehending Representational Associations [0.6554326244334868]
We show that an approach to label intermediate representations in deep neural networks fails to capture valuable information about their behaviour.
We hypothesise that non-extremal level activations contain complex information worth investigating.
We use our findings to develop a method to curate data from mid-range logit samples for retraining to mitigate spurious correlations.
arXiv Detail & Related papers (2024-11-15T07:54:14Z) - Towards Scalable and Versatile Weight Space Learning [51.78426981947659]
This paper introduces the SANE approach to weight-space learning.
Our method extends the idea of hyper-representations towards sequential processing of subsets of neural network weights.
arXiv Detail & Related papers (2024-06-14T13:12:07Z) - Identifying Interpretable Visual Features in Artificial and Biological
Neural Systems [3.604033202771937]
Single neurons in neural networks are often interpretable in that they represent individual, intuitively meaningful features.
Many neurons exhibit $textitmixed selectivity$, i.e., they represent multiple unrelated features.
We propose an automated method for quantifying visual interpretability and an approach for finding meaningful directions in network activation space.
arXiv Detail & Related papers (2023-10-17T17:41:28Z) - DISCOVER: Making Vision Networks Interpretable via Competition and
Dissection [11.028520416752325]
This work contributes to post-hoc interpretability, and specifically Network Dissection.
Our goal is to present a framework that makes it easier to discover the individual functionality of each neuron in a network trained on a vision task.
arXiv Detail & Related papers (2023-10-07T21:57:23Z) - Synergistic information supports modality integration and flexible
learning in neural networks solving multiple tasks [107.8565143456161]
We investigate the information processing strategies adopted by simple artificial neural networks performing a variety of cognitive tasks.
Results show that synergy increases as neural networks learn multiple diverse tasks.
randomly turning off neurons during training through dropout increases network redundancy, corresponding to an increase in robustness.
arXiv Detail & Related papers (2022-10-06T15:36:27Z) - A Measure of the Complexity of Neural Representations based on Partial
Information Decomposition [0.0]
In neural networks, task-relevant information is represented jointly by groups of neurons.
We show how Partial Information Decomposition (PID), a recent extension of information theory, can disentangle these different contributions.
We introduce the measure of "Representational Complexity", which quantifies the difficulty of accessing information spread across multiple neurons.
arXiv Detail & Related papers (2022-09-21T15:33:11Z) - Redundant representations help generalization in wide neural networks [71.38860635025907]
We study the last hidden layer representations of various state-of-the-art convolutional neural networks.
We find that if the last hidden representation is wide enough, its neurons tend to split into groups that carry identical information, and differ from each other only by statistically independent noise.
arXiv Detail & Related papers (2021-06-07T10:18:54Z) - Domain Generalization for Medical Imaging Classification with
Linear-Dependency Regularization [59.5104563755095]
We introduce a simple but effective approach to improve the generalization capability of deep neural networks in the field of medical imaging classification.
Motivated by the observation that the domain variability of the medical images is to some extent compact, we propose to learn a representative feature space through variational encoding.
arXiv Detail & Related papers (2020-09-27T12:30:30Z) - Exploiting Heterogeneity in Operational Neural Networks by Synaptic
Plasticity [87.32169414230822]
Recently proposed network model, Operational Neural Networks (ONNs), can generalize the conventional Convolutional Neural Networks (CNNs)
In this study the focus is drawn on searching the best-possible operator set(s) for the hidden neurons of the network based on the Synaptic Plasticity paradigm that poses the essential learning theory in biological neurons.
Experimental results over highly challenging problems demonstrate that the elite ONNs even with few neurons and layers can achieve a superior learning performance than GIS-based ONNs.
arXiv Detail & Related papers (2020-08-21T19:03:23Z) - Compositional Explanations of Neurons [52.71742655312625]
We describe a procedure for explaining neurons in deep representations by identifying compositional logical concepts.
We use this procedure to answer several questions on interpretability in models for vision and natural language processing.
arXiv Detail & Related papers (2020-06-24T20:37:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.