Structural Learning of Simple Staged Trees
- URL: http://arxiv.org/abs/2203.04390v1
- Date: Tue, 8 Mar 2022 20:50:39 GMT
- Title: Structural Learning of Simple Staged Trees
- Authors: Manuele Leonelli and Gherardo Varando
- Abstract summary: We introduce the first structural learning algorithms for the class of simple staged trees.
We show that data-learned simple staged trees often outperform Bayesian networks in model fit.
- Score: 2.3572498744567127
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Bayesian networks faithfully represent the symmetric conditional
independences existing between the components of a random vector. Staged trees
are an extension of Bayesian networks for categorical random vectors whose
graph represents non-symmetric conditional independences via vertex coloring.
However, since they are based on a tree representation of the sample space, the
underlying graph becomes cluttered and difficult to visualize as the number of
variables increases. Here we introduce the first structural learning algorithms
for the class of simple staged trees, entertaining a compact coalescence of the
underlying tree from which non-symmetric independences can be easily read. We
show that data-learned simple staged trees often outperform Bayesian networks
in model fit and illustrate how the coalesced graph is used to identify
non-symmetric conditional independences.
Related papers
- Learning Staged Trees from Incomplete Data [1.6327794667678908]
We introduce the first algorithms for staged trees that handle missingness within the learning of the model.
A computational experiment showcases the performance of the novel learning algorithms.
arXiv Detail & Related papers (2024-05-28T16:00:23Z) - Highly Efficient Structural Learning of Sparse Staged Trees [2.3572498744567127]
We introduce the first scalable structural learning algorithm for staged trees, which searches over a space of models where only a small number of dependencies can be imposed.
A simulation study as well as a real-world application illustrate our routines and the practical use of such data-learned staged trees.
arXiv Detail & Related papers (2022-06-14T16:46:13Z) - Active-LATHE: An Active Learning Algorithm for Boosting the Error
Exponent for Learning Homogeneous Ising Trees [75.93186954061943]
We design and analyze an algorithm that boosts the error exponent by at least 40% when $rho$ is at least $0.8$.
Our analysis hinges on judiciously exploiting the minute but detectable statistical variation of the samples to allocate more data to parts of the graph.
arXiv Detail & Related papers (2021-10-27T10:45:21Z) - Staged trees and asymmetry-labeled DAGs [2.66269503676104]
We introduce a minimal Bayesian network representation of the staged tree, which can be used to read conditional independences in an intuitive way.
We also define a new labeled graph, termed asymmetry-labeled directed acyclic graph, whose edges are labeled to denote the type of dependence existing between any two random variables.
arXiv Detail & Related papers (2021-08-04T12:20:47Z) - Learning compositional structures for semantic graph parsing [81.41592892863979]
We show how AM dependency parsing can be trained directly on a neural latent-variable model.
Our model picks up on several linguistic phenomena on its own and achieves comparable accuracy to supervised training.
arXiv Detail & Related papers (2021-06-08T14:20:07Z) - Spectral Top-Down Recovery of Latent Tree Models [13.681975313065477]
Spectral Top-Down Recovery (STDR) is a divide-and-conquer approach for inference of large latent tree models.
STDR's partitioning step is non-random. Instead, it is based on the Fiedler vector of a suitable Laplacian matrix related to the observed nodes.
We prove that STDR is statistically consistent, and bound the number of samples required to accurately recover the tree with high probability.
arXiv Detail & Related papers (2021-02-26T02:47:42Z) - Visualizing hierarchies in scRNA-seq data using a density tree-biased
autoencoder [50.591267188664666]
We propose an approach for identifying a meaningful tree structure from high-dimensional scRNA-seq data.
We then introduce DTAE, a tree-biased autoencoder that emphasizes the tree structure of the data in low dimensional space.
arXiv Detail & Related papers (2021-02-11T08:48:48Z) - SGA: A Robust Algorithm for Partial Recovery of Tree-Structured
Graphical Models with Noisy Samples [75.32013242448151]
We consider learning Ising tree models when the observations from the nodes are corrupted by independent but non-identically distributed noise.
Katiyar et al. (2020) showed that although the exact tree structure cannot be recovered, one can recover a partial tree structure.
We propose Symmetrized Geometric Averaging (SGA), a more statistically robust algorithm for partial tree recovery.
arXiv Detail & Related papers (2021-01-22T01:57:35Z) - Predicting What You Already Know Helps: Provable Self-Supervised
Learning [60.27658820909876]
Self-supervised representation learning solves auxiliary prediction tasks (known as pretext tasks) without requiring labeled data.
We show a mechanism exploiting the statistical connections between certain em reconstruction-based pretext tasks that guarantee to learn a good representation.
We prove the linear layer yields small approximation error even for complex ground truth function class.
arXiv Detail & Related papers (2020-08-03T17:56:13Z) - Constructing a Chain Event Graph from a Staged Tree [0.0]
Chain Event Graphs (CEGs) are a recent family of probabilistic graphical models.
No general algorithm has yet been devised that automatically transforms any staged tree into a CEG representation.
We show that no information is lost from transforming a staged tree into a CEG.
arXiv Detail & Related papers (2020-06-29T08:07:06Z) - Linguistically Driven Graph Capsule Network for Visual Question
Reasoning [153.76012414126643]
We propose a hierarchical compositional reasoning model called the "Linguistically driven Graph Capsule Network"
The compositional process is guided by the linguistic parse tree. Specifically, we bind each capsule in the lowest layer to bridge the linguistic embedding of a single word in the original question with visual evidence.
Experiments on the CLEVR dataset, CLEVR compositional generation test, and FigureQA dataset demonstrate the effectiveness and composition generalization ability of our end-to-end model.
arXiv Detail & Related papers (2020-03-23T03:34:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.