Similarity of Neural Networks with Gradients
- URL: http://arxiv.org/abs/2003.11498v1
- Date: Wed, 25 Mar 2020 17:04:10 GMT
- Title: Similarity of Neural Networks with Gradients
- Authors: Shuai Tang, Wesley J. Maddox, Charlie Dickens, Tom Diethe, Andreas
Damianou
- Abstract summary: We propose to leverage both feature vectors and gradient ones into designing the representation of a neural network.
We show that the proposed approach provides a state-of-the-art method for computing similarity of neural networks.
- Score: 8.804507286438781
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A suitable similarity index for comparing learnt neural networks plays an
important role in understanding the behaviour of the highly-nonlinear
functions, and can provide insights on further theoretical analysis and
empirical studies. We define two key steps when comparing models: firstly, the
representation abstracted from the learnt model, where we propose to leverage
both feature vectors and gradient ones (which are largely ignored in prior
work) into designing the representation of a neural network. Secondly, we
define the employed similarity index which gives desired invariance properties,
and we facilitate the chosen ones with sketching techniques for comparing
various datasets efficiently. Empirically, we show that the proposed approach
provides a state-of-the-art method for computing similarity of neural networks
that are trained independently on different datasets and the tasks defined by
the datasets.
Related papers
- Graph Neural Networks for Learning Equivariant Representations of Neural Networks [55.04145324152541]
We propose to represent neural networks as computational graphs of parameters.
Our approach enables a single model to encode neural computational graphs with diverse architectures.
We showcase the effectiveness of our method on a wide range of tasks, including classification and editing of implicit neural representations.
arXiv Detail & Related papers (2024-03-18T18:01:01Z) - Demolition and Reinforcement of Memories in Spin-Glass-like Neural
Networks [0.0]
The aim of this thesis is to understand the effectiveness of Unlearning in both associative memory models and generative models.
The selection of structured data enables an associative memory model to retrieve concepts as attractors of a neural dynamics with considerable basins of attraction.
A novel regularization technique for Boltzmann Machines is presented, proving to outperform previously developed methods in learning hidden probability distributions from data-sets.
arXiv Detail & Related papers (2024-03-04T23:12:42Z) - Multilayer Multiset Neuronal Networks -- MMNNs [55.2480439325792]
The present work describes multilayer multiset neuronal networks incorporating two or more layers of coincidence similarity neurons.
The work also explores the utilization of counter-prototype points, which are assigned to the image regions to be avoided.
arXiv Detail & Related papers (2023-08-28T12:55:13Z) - Object-based Probabilistic Similarity Evidence of Sparse Latent Features
from Fully Convolutional Networks [0.0]
Similarity analysis using neural networks has emerged as a powerful technique for understanding and categorizing complex patterns in various domains.
This research explores the utilization of latent information generated by fully convolutional networks (FCNs) in similarity analysis.
arXiv Detail & Related papers (2023-07-25T16:15:29Z) - Do Neural Networks Trained with Topological Features Learn Different
Internal Representations? [1.418465438044804]
We investigate whether a model trained with topological features learns internal representations of data that are fundamentally different than those learned by a model trained with the original raw data.
We find that structurally, the hidden representations of models trained and evaluated on topological features differ substantially compared to those trained and evaluated on the corresponding raw data.
We conjecture that this means that neural networks trained on raw data may extract some limited topological features in the process of making predictions.
arXiv Detail & Related papers (2022-11-14T19:19:04Z) - Data-driven emergence of convolutional structure in neural networks [83.4920717252233]
We show how fully-connected neural networks solving a discrimination task can learn a convolutional structure directly from their inputs.
By carefully designing data models, we show that the emergence of this pattern is triggered by the non-Gaussian, higher-order local structure of the inputs.
arXiv Detail & Related papers (2022-02-01T17:11:13Z) - Similarity and Matching of Neural Network Representations [0.0]
We employ a toolset -- dubbed Dr. Frankenstein -- to analyse the similarity of representations in deep neural networks.
We aim to match the activations on given layers of two trained neural networks by joining them with a stitching layer.
arXiv Detail & Related papers (2021-10-27T17:59:46Z) - Discriminative Attribution from Counterfactuals [64.94009515033984]
We present a method for neural network interpretability by combining feature attribution with counterfactual explanations.
We show that this method can be used to quantitatively evaluate the performance of feature attribution methods in an objective manner.
arXiv Detail & Related papers (2021-09-28T00:53:34Z) - Deep Representational Similarity Learning for analyzing neural
signatures in task-based fMRI dataset [81.02949933048332]
This paper develops Deep Representational Similarity Learning (DRSL), a deep extension of Representational Similarity Analysis (RSA)
DRSL is appropriate for analyzing similarities between various cognitive tasks in fMRI datasets with a large number of subjects.
arXiv Detail & Related papers (2020-09-28T18:30:14Z) - Learning What Makes a Difference from Counterfactual Examples and
Gradient Supervision [57.14468881854616]
We propose an auxiliary training objective that improves the generalization capabilities of neural networks.
We use pairs of minimally-different examples with different labels, a.k.a counterfactual or contrasting examples, which provide a signal indicative of the underlying causal structure of the task.
Models trained with this technique demonstrate improved performance on out-of-distribution test sets.
arXiv Detail & Related papers (2020-04-20T02:47:49Z) - Gradients as Features for Deep Representation Learning [26.996104074384263]
We address the problem of deep representation learning--the efficient adaption of a pre-trained deep network to different tasks.
Our key innovation is the design of a linear model that incorporates both gradient and activation of the pre-trained network.
We present an efficient algorithm for the training and inference of our model without computing the actual gradient.
arXiv Detail & Related papers (2020-04-12T02:57:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.