Approaching an unknown communication system by latent space exploration
and causal inference
- URL: http://arxiv.org/abs/2303.10931v2
- Date: Tue, 6 Feb 2024 14:05:49 GMT
- Title: Approaching an unknown communication system by latent space exploration
and causal inference
- Authors: Ga\v{s}per Begu\v{s} and Andrej Leban, Shane Gero
- Abstract summary: This paper proposes a methodology for discovering meaningful properties in data by exploring the latent space of unsupervised deep generative models.
We combine manipulation of individual latent variables to extreme values with methods inspired by causal inference.
We show that this method yields insights for model interpretability.
- Score: 5.026037329977691
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper proposes a methodology for discovering meaningful properties in
data by exploring the latent space of unsupervised deep generative models. We
combine manipulation of individual latent variables to extreme values with
methods inspired by causal inference into an approach we call causal
disentanglement with extreme values (CDEV) and show that this method yields
insights for model interpretability. With this, we can test for what properties
of unknown data the model encodes as meaningful, using it to glean insight into
the communication system of sperm whales (Physeter macrocephalus), one of the
most intriguing and understudied animal communication systems. The network
architecture used has been shown to learn meaningful representations of speech;
here, it is used as a learning mechanism to decipher the properties of another
vocal communication system in which case we have no ground truth. The proposed
methodology suggests that sperm whales encode information using the number of
clicks in a sequence, the regularity of their timing, and audio properties such
as the spectral mean and the acoustic regularity of the sequences. Some of
these findings are consistent with existing hypotheses, while others are
proposed for the first time. We also argue that our models uncover rules that
govern the structure of units in the communication system and apply them while
generating innovative data not shown during training. This paper suggests that
an interpretation of the outputs of deep neural networks with causal inference
methodology can be a viable strategy for approaching data about which little is
known and presents another case of how deep learning can limit the hypothesis
space. Finally, the proposed approach can be extended to other architectures
and datasets.
Related papers
- Seeing Unseen: Discover Novel Biomedical Concepts via
Geometry-Constrained Probabilistic Modeling [53.7117640028211]
We present a geometry-constrained probabilistic modeling treatment to resolve the identified issues.
We incorporate a suite of critical geometric properties to impose proper constraints on the layout of constructed embedding space.
A spectral graph-theoretic method is devised to estimate the number of potential novel classes.
arXiv Detail & Related papers (2024-03-02T00:56:05Z) - Prototype-based Aleatoric Uncertainty Quantification for Cross-modal
Retrieval [139.21955930418815]
Cross-modal Retrieval methods build similarity relations between vision and language modalities by jointly learning a common representation space.
However, the predictions are often unreliable due to the Aleatoric uncertainty, which is induced by low-quality data, e.g., corrupt images, fast-paced videos, and non-detailed texts.
We propose a novel Prototype-based Aleatoric Uncertainty Quantification (PAU) framework to provide trustworthy predictions by quantifying the uncertainty arisen from the inherent data ambiguity.
arXiv Detail & Related papers (2023-09-29T09:41:19Z) - On the Joint Interaction of Models, Data, and Features [82.60073661644435]
We introduce a new tool, the interaction tensor, for empirically analyzing the interaction between data and model through features.
Based on these observations, we propose a conceptual framework for feature learning.
Under this framework, the expected accuracy for a single hypothesis and agreement for a pair of hypotheses can both be derived in closed-form.
arXiv Detail & Related papers (2023-06-07T21:35:26Z) - Deep networks for system identification: a Survey [56.34005280792013]
System identification learns mathematical descriptions of dynamic systems from input-output data.
Main aim of the identified model is to predict new data from previous observations.
We discuss architectures commonly adopted in the literature, like feedforward, convolutional, and recurrent networks.
arXiv Detail & Related papers (2023-01-30T12:38:31Z) - Experimental Observations of the Topology of Convolutional Neural
Network Activations [2.4235626091331737]
Topological data analysis provides compact, noise-robust representations of complex structures.
Deep neural networks (DNNs) learn millions of parameters associated with a series of transformations defined by the model architecture.
In this paper, we apply cutting edge techniques from TDA with the goal of gaining insight into the interpretability of convolutional neural networks used for image classification.
arXiv Detail & Related papers (2022-12-01T02:05:44Z) - Learning Dynamics from Noisy Measurements using Deep Learning with a
Runge-Kutta Constraint [9.36739413306697]
We discuss a methodology to learn differential equation(s) using noisy and sparsely sampled measurements.
In our methodology, the main innovation can be seen in of integration of deep neural networks with a classical numerical integration method.
arXiv Detail & Related papers (2021-09-23T15:43:45Z) - The Causal Neural Connection: Expressiveness, Learnability, and
Inference [125.57815987218756]
An object called structural causal model (SCM) represents a collection of mechanisms and sources of random variation of the system under investigation.
In this paper, we show that the causal hierarchy theorem (Thm. 1, Bareinboim et al., 2020) still holds for neural models.
We introduce a special type of SCM called a neural causal model (NCM), and formalize a new type of inductive bias to encode structural constraints necessary for performing causal inferences.
arXiv Detail & Related papers (2021-07-02T01:55:18Z) - Model discovery in the sparse sampling regime [0.0]
We show how deep learning can improve model discovery of partial differential equations.
As a result, deep learning-based model discovery allows to recover the underlying equations.
We illustrate our claims on both synthetic and experimental sets.
arXiv Detail & Related papers (2021-05-02T06:27:05Z) - Deep Co-Attention Network for Multi-View Subspace Learning [73.3450258002607]
We propose a deep co-attention network for multi-view subspace learning.
It aims to extract both the common information and the complementary information in an adversarial setting.
In particular, it uses a novel cross reconstruction loss and leverages the label information to guide the construction of the latent representation.
arXiv Detail & Related papers (2021-02-15T18:46:44Z) - Local and non-local dependency learning and emergence of rule-like
representations in speech data by Deep Convolutional Generative Adversarial
Networks [0.0]
This paper argues that training GANs on local and non-local dependencies in speech data offers insights into how deep neural networks discretize continuous data.
arXiv Detail & Related papers (2020-09-27T00:02:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.