Invariant Representations with Stochastically Quantized Neural Networks
- URL: http://arxiv.org/abs/2208.02656v1
- Date: Thu, 4 Aug 2022 13:36:06 GMT
- Title: Invariant Representations with Stochastically Quantized Neural Networks
- Authors: Mattia Cerrato, Marius K\"oppel, Roberto Esposito, Stefan Kramer
- Abstract summary: We propose a methodology for direct computation of the mutual information between a neural layer and a sensitive attribute.
We show that this method compares favorably with the state of the art in fair representation learning.
- Score: 5.7923858184309385
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Representation learning algorithms offer the opportunity to learn invariant
representations of the input data with regard to nuisance factors. Many authors
have leveraged such strategies to learn fair representations, i.e., vectors
where information about sensitive attributes is removed. These methods are
attractive as they may be interpreted as minimizing the mutual information
between a neural layer's activations and a sensitive attribute. However, the
theoretical grounding of such methods relies either on the computation of
infinitely accurate adversaries or on minimizing a variational upper bound of a
mutual information estimate. In this paper, we propose a methodology for direct
computation of the mutual information between a neural layer and a sensitive
attribute. We employ stochastically-activated binary neural networks, which
lets us treat neurons as random variables. We are then able to compute (not
bound) the mutual information between a layer and a sensitive attribute and use
this information as a regularization factor during gradient descent. We show
that this method compares favorably with the state of the art in fair
representation learning and that the learned representations display a higher
level of invariance compared to full-precision neural networks.
Related papers
- Graph Neural Networks for Learning Equivariant Representations of Neural Networks [55.04145324152541]
We propose to represent neural networks as computational graphs of parameters.
Our approach enables a single model to encode neural computational graphs with diverse architectures.
We showcase the effectiveness of our method on a wide range of tasks, including classification and editing of implicit neural representations.
arXiv Detail & Related papers (2024-03-18T18:01:01Z) - Manipulating Feature Visualizations with Gradient Slingshots [54.31109240020007]
We introduce a novel method for manipulating Feature Visualization (FV) without significantly impacting the model's decision-making process.
We evaluate the effectiveness of our method on several neural network models and demonstrate its capabilities to hide the functionality of arbitrarily chosen neurons.
arXiv Detail & Related papers (2024-01-11T18:57:17Z) - Generalizable Neural Fields as Partially Observed Neural Processes [16.202109517569145]
We propose a new paradigm that views the large-scale training of neural representations as a part of a partially-observed neural process framework.
We demonstrate that this approach outperforms both state-of-the-art gradient-based meta-learning approaches and hypernetwork approaches.
arXiv Detail & Related papers (2023-09-13T01:22:16Z) - Unsupervised Learning of Invariance Transformations [105.54048699217668]
We develop an algorithmic framework for finding approximate graph automorphisms.
We discuss how this framework can be used to find approximate automorphisms in weighted graphs in general.
arXiv Detail & Related papers (2023-07-24T17:03:28Z) - Representational dissimilarity metric spaces for stochastic neural
networks [4.229248343585332]
Quantifying similarity between neural representations is a perennial problem in deep learning and neuroscience research.
We generalize shape metrics to quantify differences in representations.
We find that neurobiological oriented visual gratings and naturalistic scenes respectively resemble untrained and trained deep network representations.
arXiv Detail & Related papers (2022-11-21T17:32:40Z) - Data-driven emergence of convolutional structure in neural networks [83.4920717252233]
We show how fully-connected neural networks solving a discrimination task can learn a convolutional structure directly from their inputs.
By carefully designing data models, we show that the emergence of this pattern is triggered by the non-Gaussian, higher-order local structure of the inputs.
arXiv Detail & Related papers (2022-02-01T17:11:13Z) - Disentangled representations: towards interpretation of sex
determination from hip bone [1.0775419935941009]
saliency maps have become a popular method to make neural networks interpretable.
We propose a new paradigm for better interpretability.
We illustrate the relevance of this approach in the context of automatic sex determination from hip bones in forensic medicine.
arXiv Detail & Related papers (2021-12-17T10:07:05Z) - Drop, Swap, and Generate: A Self-Supervised Approach for Generating
Neural Activity [33.06823702945747]
We introduce a novel unsupervised approach for learning disentangled representations of neural activity called Swap-VAE.
Our approach combines a generative modeling framework with an instance-specific alignment loss.
We show that it is possible to build representations that disentangle neural datasets along relevant latent dimensions linked to behavior.
arXiv Detail & Related papers (2021-11-03T16:39:43Z) - Discriminative Attribution from Counterfactuals [64.94009515033984]
We present a method for neural network interpretability by combining feature attribution with counterfactual explanations.
We show that this method can be used to quantitatively evaluate the performance of feature attribution methods in an objective manner.
arXiv Detail & Related papers (2021-09-28T00:53:34Z) - Category-Learning with Context-Augmented Autoencoder [63.05016513788047]
Finding an interpretable non-redundant representation of real-world data is one of the key problems in Machine Learning.
We propose a novel method of using data augmentations when training autoencoders.
We train a Variational Autoencoder in such a way, that it makes transformation outcome predictable by auxiliary network.
arXiv Detail & Related papers (2020-10-10T14:04:44Z) - Malicious Network Traffic Detection via Deep Learning: An Information
Theoretic View [0.0]
We study how homeomorphism affects learned representation of a malware traffic dataset.
Our results suggest that although the details of learned representations and the specific coordinate system defined over the manifold of all parameters differ slightly, the functional approximations are the same.
arXiv Detail & Related papers (2020-09-16T15:37:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.