Multi-view hierarchical Variational AutoEncoders with Factor Analysis
latent space
- URL: http://arxiv.org/abs/2207.09185v1
- Date: Tue, 19 Jul 2022 10:46:02 GMT
- Title: Multi-view hierarchical Variational AutoEncoders with Factor Analysis
latent space
- Authors: Alejandro Guerrero-L\'opez, Carlos Sevilla-Salcedo, Vanessa
G\'omez-Verdejo, Pablo M. Olmos
- Abstract summary: We propose a novel method to combine multiple Variational AutoEncoders with a Factor Analysis latent space.
We create an interpretable hierarchical dependency between private and shared information.
This way, the novel model is able to simultaneously: (i) learn from multiple heterogeneous views, (ii) obtain an interpretable hierarchical shared space, and (iii) perform transfer learning between generative models.
- Score: 67.60224656603823
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Real-world databases are complex, they usually present redundancy and shared
correlations between heterogeneous and multiple representations of the same
data. Thus, exploiting and disentangling shared information between views is
critical. For this purpose, recent studies often fuse all views into a shared
nonlinear complex latent space but they lose the interpretability. To overcome
this limitation, here we propose a novel method to combine multiple Variational
AutoEncoders (VAE) architectures with a Factor Analysis latent space (FA-VAE).
Concretely, we use a VAE to learn a private representation of each
heterogeneous view in a continuous latent space. Then, we model the shared
latent space by projecting every private variable to a low-dimensional latent
space using a linear projection matrix. Thus, we create an interpretable
hierarchical dependency between private and shared information. This way, the
novel model is able to simultaneously: (i) learn from multiple heterogeneous
views, (ii) obtain an interpretable hierarchical shared space, and, (iii)
perform transfer learning between generative models.
Related papers
- Diversity as a Reward: Fine-Tuning LLMs on a Mixture of Domain-Undetermined Data [36.277423093218275]
We study the role of data diversity in enhancing the overall abilities of large language models (LLMs)
We propose a new method that gives the LLM a dual identity: an output model to cognitively probe and select data based on diversity reward, as well as an input model to be tuned with the selected data.
arXiv Detail & Related papers (2025-02-05T17:21:01Z) - Learning Divergence Fields for Shift-Robust Graph Representations [73.11818515795761]
In this work, we propose a geometric diffusion model with learnable divergence fields for the challenging problem with interdependent data.
We derive a new learning objective through causal inference, which can guide the model to learn generalizable patterns of interdependence that are insensitive across domains.
arXiv Detail & Related papers (2024-06-07T14:29:21Z) - Flexible inference in heterogeneous and attributed multilayer networks [21.349513661012498]
We develop a probabilistic generative model to perform inference in multilayer networks with arbitrary types of information.
We demonstrate its ability to unveil a variety of patterns in a social support network among villagers in rural India.
arXiv Detail & Related papers (2024-05-31T15:21:59Z) - An improved tabular data generator with VAE-GMM integration [9.4491536689161]
We propose a novel Variational Autoencoder (VAE)-based model that addresses limitations of current approaches.
Inspired by the TVAE model, our approach incorporates a Bayesian Gaussian Mixture model (BGM) within the VAE architecture.
We thoroughly validate our model on three real-world datasets with mixed data types, including two medically relevant ones.
arXiv Detail & Related papers (2024-04-12T12:31:06Z) - VTAE: Variational Transformer Autoencoder with Manifolds Learning [144.0546653941249]
Deep generative models have demonstrated successful applications in learning non-linear data distributions through a number of latent variables.
The nonlinearity of the generator implies that the latent space shows an unsatisfactory projection of the data space, which results in poor representation learning.
We show that geodesics and accurate computation can substantially improve the performance of deep generative models.
arXiv Detail & Related papers (2023-04-03T13:13:19Z) - Dynamic Latent Separation for Deep Learning [67.62190501599176]
A core problem in machine learning is to learn expressive latent variables for model prediction on complex data.
Here, we develop an approach that improves expressiveness, provides partial interpretation, and is not restricted to specific applications.
arXiv Detail & Related papers (2022-10-07T17:56:53Z) - Encoding Domain Knowledge in Multi-view Latent Variable Models: A
Bayesian Approach with Structured Sparsity [7.811916700683125]
MuVI is a novel approach for domain-informed multi-view latent variable models.
We demonstrate that our model is able to integrate noisy domain expertise in form of feature sets.
arXiv Detail & Related papers (2022-04-13T08:22:31Z) - Heterogeneous Target Speech Separation [52.05046029743995]
We introduce a new paradigm for single-channel target source separation where the sources of interest can be distinguished using non-mutually exclusive concepts.
Our proposed heterogeneous separation framework can seamlessly leverage datasets with large distribution shifts.
arXiv Detail & Related papers (2022-04-07T17:14:20Z) - MoReL: Multi-omics Relational Learning [26.484803417186384]
We propose a novel deep Bayesian generative model to efficiently infer a multi-partite graph encoding molecular interactions across heterogeneous views.
With such an optimal transport regularization in the deep Bayesian generative model, it not only allows incorporating view-specific side information, but also increases the model flexibility with the distribution-based regularization.
arXiv Detail & Related papers (2022-03-15T02:50:07Z) - Consistency and Diversity induced Human Motion Segmentation [231.36289425663702]
We propose a novel Consistency and Diversity induced human Motion (CDMS) algorithm.
Our model factorizes the source and target data into distinct multi-layer feature spaces.
A multi-mutual learning strategy is carried out to reduce the domain gap between the source and target data.
arXiv Detail & Related papers (2022-02-10T06:23:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.