Deep Probabilistic Graphical Modeling
- URL: http://arxiv.org/abs/2104.12053v1
- Date: Sun, 25 Apr 2021 03:48:02 GMT
- Title: Deep Probabilistic Graphical Modeling
- Authors: Adji B. Dieng
- Abstract summary: This thesis develops deep probabilistic graphical modeling (DPGM)
DPGM consists in leveraging deep learning (DL) to make PGM more flexible.
One model class we develop extends exponential family PCA using neural networks to improve predictive performance.
- Score: 2.2691593216516863
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Probabilistic graphical modeling (PGM) provides a framework for formulating
an interpretable generative process of data and expressing uncertainty about
unknowns, but it lacks flexibility. Deep learning (DL) is an alternative
framework for learning from data that has achieved great empirical success in
recent years. DL offers great flexibility, but it lacks the interpretability
and calibration of PGM. This thesis develops deep probabilistic graphical
modeling (DPGM.) DPGM consists in leveraging DL to make PGM more flexible. DPGM
brings about new methods for learning from data that exhibit the advantages of
both PGM and DL.
We use DL within PGM to build flexible models endowed with an interpretable
latent structure. One model class we develop extends exponential family PCA
using neural networks to improve predictive performance while enforcing the
interpretability of the latent factors. Another model class we introduce
enables accounting for long-term dependencies when modeling sequential data,
which is a challenge when using purely DL or PGM approaches. Finally, DPGM
successfully solves several outstanding problems of probabilistic topic models,
a widely used family of models in PGM.
DPGM also brings about new algorithms for learning with complex data. We
develop reweighted expectation maximization, an algorithm that unifies several
existing maximum likelihood-based algorithms for learning models parameterized
by neural networks. This unifying view is made possible using expectation
maximization, a canonical inference algorithm in PGM. We also develop
entropy-regularized adversarial learning, a learning paradigm that deviates
from the traditional maximum likelihood approach used in PGM. From the DL
perspective, entropy-regularized adversarial learning provides a solution to
the long-standing mode collapse problem of generative adversarial networks, a
widely used DL approach.
Related papers
- Sample Complexity Characterization for Linear Contextual MDPs [67.79455646673762]
Contextual decision processes (CMDPs) describe a class of reinforcement learning problems in which the transition kernels and reward functions can change over time with different MDPs indexed by a context variable.
CMDPs serve as an important framework to model many real-world applications with time-varying environments.
We study CMDPs under two linear function approximation models: Model I with context-varying representations and common linear weights for all contexts; and Model II with common representations for all contexts and context-varying linear weights.
arXiv Detail & Related papers (2024-02-05T03:25:04Z) - Predictable MDP Abstraction for Unsupervised Model-Based RL [93.91375268580806]
We propose predictable MDP abstraction (PMA)
Instead of training a predictive model on the original MDP, we train a model on a transformed MDP with a learned action space.
We theoretically analyze PMA and empirically demonstrate that PMA leads to significant improvements over prior unsupervised model-based RL approaches.
arXiv Detail & Related papers (2023-02-08T07:37:51Z) - PAC Reinforcement Learning for Predictive State Representations [60.00237613646686]
We study online Reinforcement Learning (RL) in partially observable dynamical systems.
We focus on the Predictive State Representations (PSRs) model, which is an expressive model that captures other well-known models.
We develop a novel model-based algorithm for PSRs that can learn a near optimal policy in sample complexity scalingly.
arXiv Detail & Related papers (2022-07-12T17:57:17Z) - Bayesian Active Learning for Discrete Latent Variable Models [19.852463786440122]
Active learning seeks to reduce the amount of data required to fit the parameters of a model.
latent variable models play a vital role in neuroscience, psychology, and a variety of other engineering and scientific disciplines.
arXiv Detail & Related papers (2022-02-27T19:07:12Z) - Modeling Item Response Theory with Stochastic Variational Inference [8.369065078321215]
We introduce a variational Bayesian inference algorithm for Item Response Theory (IRT)
Applying this method to five large-scale item response datasets yields higher log likelihoods and higher accuracy in imputing missing data.
The algorithm implementation is open-source, and easily usable.
arXiv Detail & Related papers (2021-08-26T05:00:27Z) - Closed-form Continuous-Depth Models [99.40335716948101]
Continuous-depth neural models rely on advanced numerical differential equation solvers.
We present a new family of models, termed Closed-form Continuous-depth (CfC) networks, that are simple to describe and at least one order of magnitude faster.
arXiv Detail & Related papers (2021-06-25T22:08:51Z) - A Deep Latent Space Model for Graph Representation Learning [10.914558012458425]
We propose a Deep Latent Space Model (DLSM) for directed graphs to incorporate the traditional latent variable based generative model into deep learning frameworks.
Our proposed model consists of a graph convolutional network (GCN) encoder and a decoder, which are layer-wise connected by a hierarchical variational auto-encoder architecture.
Experiments on real-world datasets show that the proposed model achieves the state-of-the-art performances on both link prediction and community detection tasks.
arXiv Detail & Related papers (2021-06-22T12:41:19Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z) - Deep neural network enabled corrective source term approach to hybrid
analysis and modeling [0.0]
Hybrid Analysis and Modeling (HAM) is an emerging modeling paradigm which aims to combine physics-based modeling and data-driven modeling.
We introduce, justify and demonstrate a novel approach to HAM -- the Corrective Source Term Approach (CoSTA)
arXiv Detail & Related papers (2021-05-24T20:17:13Z) - Improving the Reconstruction of Disentangled Representation Learners via Multi-Stage Modeling [54.94763543386523]
Current autoencoder-based disentangled representation learning methods achieve disentanglement by penalizing the ( aggregate) posterior to encourage statistical independence of the latent factors.
We present a novel multi-stage modeling approach where the disentangled factors are first learned using a penalty-based disentangled representation learning method.
Then, the low-quality reconstruction is improved with another deep generative model that is trained to model the missing correlated latent variables.
arXiv Detail & Related papers (2020-10-25T18:51:15Z) - Causality-aware counterfactual confounding adjustment for feature
representations learned by deep models [14.554818659491644]
Causal modeling has been recognized as a potential solution to many challenging problems in machine learning (ML)
We describe how a recently proposed counterfactual approach can still be used to deconfound the feature representations learned by deep neural network (DNN) models.
arXiv Detail & Related papers (2020-04-20T17:37:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.