A Bayesian Hierarchical Score for Structure Learning from Related Data
Sets
- URL: http://arxiv.org/abs/2008.01683v3
- Date: Sat, 17 Jul 2021 16:32:29 GMT
- Title: A Bayesian Hierarchical Score for Structure Learning from Related Data
Sets
- Authors: Laura Azzimonti, Giorgio Corani and Marco Scutari
- Abstract summary: We propose a new Bayesian Dirichlet score, which we call Bayesian Hierarchical Dirichlet (BHD)
BHD is based on a hierarchical model that pools information across data sets to learn a single encompassing network structure.
We find that BHD outperforms the Bayesian Dirichlet equivalent uniform (BDeu) score in terms of reconstruction accuracy as measured by the Structural Hamming distance.
- Score: 0.7240563090941907
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Score functions for learning the structure of Bayesian networks in the
literature assume that data are a homogeneous set of observations; whereas it
is often the case that they comprise different related, but not homogeneous,
data sets collected in different ways. In this paper we propose a new Bayesian
Dirichlet score, which we call Bayesian Hierarchical Dirichlet (BHD). The
proposed score is based on a hierarchical model that pools information across
data sets to learn a single encompassing network structure, while taking into
account the differences in their probabilistic structures. We derive a
closed-form expression for BHD using a variational approximation of the
marginal likelihood, we study the associated computational cost and we evaluate
its performance using simulated data. We find that, when data comprise multiple
related data sets, BHD outperforms the Bayesian Dirichlet equivalent uniform
(BDeu) score in terms of reconstruction accuracy as measured by the Structural
Hamming distance, and that it is as accurate as BDeu when data are homogeneous.
This improvement is particularly clear when either the number of variables in
the network or the number of observations is large. Moreover, the estimated
networks are sparser and therefore more interpretable than those obtained with
BDeu thanks to a lower number of false positive arcs.
Related papers
- Amortised Inference in Neural Networks for Small-Scale Probabilistic
Meta-Learning [41.85464593920907]
A global inducing point variational approximation for BNNs is based on using a set of inducing inputs to construct a series of conditional distributions.
Our key insight is that these inducing inputs can be replaced by the actual data, such that the variational distribution consists of a set of approximate likelihoods for each datapoint.
By training this inference network across related datasets, we can meta-learn Bayesian inference over task-specific BNNs.
arXiv Detail & Related papers (2023-10-24T12:34:25Z) - Topological Quality of Subsets via Persistence Matching Diagrams [0.196629787330046]
We measure the quality of a subset concerning the dataset it represents using topological data analysis techniques.
In particular, this approach enables us to explain why the chosen subset is likely to result in poor performance of a supervised learning model.
arXiv Detail & Related papers (2023-06-04T17:08:41Z) - Score Approximation, Estimation and Distribution Recovery of Diffusion
Models on Low-Dimensional Data [68.62134204367668]
This paper studies score approximation, estimation, and distribution recovery of diffusion models, when data are supported on an unknown low-dimensional linear subspace.
We show that with a properly chosen neural network architecture, the score function can be both accurately approximated and efficiently estimated.
The generated distribution based on the estimated score function captures the data geometric structures and converges to a close vicinity of the data distribution.
arXiv Detail & Related papers (2023-02-14T17:02:35Z) - Recursive Neural Networks with Bottlenecks Diagnose
(Non-)Compositionality [65.60002535580298]
Quantifying compositionality of data is a challenging task, which has been investigated primarily for short utterances.
We show that comparing data's representations in models with and without a bottleneck can be used to produce a compositionality metric.
The procedure is applied to the evaluation of arithmetic expressions using synthetic data, and sentiment classification using natural language data.
arXiv Detail & Related papers (2023-01-31T15:46:39Z) - Concrete Score Matching: Generalized Score Matching for Discrete Data [109.12439278055213]
"Concrete score" is a generalization of the (Stein) score for discrete settings.
"Concrete Score Matching" is a framework to learn such scores from samples.
arXiv Detail & Related papers (2022-11-02T00:41:37Z) - Using Mixed-Effect Models to Learn Bayesian Networks from Related Data
Sets [0.04297070083645048]
We provide an analogous solution for learning a Bayesian network from continuous data using mixed-effects models.
We study its structural, parametric, predictive and classification accuracy.
The improvement is marked for low sample sizes and for unbalanced data sets.
arXiv Detail & Related papers (2022-06-08T08:32:32Z) - Bayesian Structure Learning with Generative Flow Networks [85.84396514570373]
In Bayesian structure learning, we are interested in inferring a distribution over the directed acyclic graph (DAG) from data.
Recently, a class of probabilistic models, called Generative Flow Networks (GFlowNets), have been introduced as a general framework for generative modeling.
We show that our approach, called DAG-GFlowNet, provides an accurate approximation of the posterior over DAGs.
arXiv Detail & Related papers (2022-02-28T15:53:10Z) - BCDAG: An R package for Bayesian structure and Causal learning of
Gaussian DAGs [77.34726150561087]
We introduce the R package for causal discovery and causal effect estimation from observational data.
Our implementation scales efficiently with the number of observations and, whenever the DAGs are sufficiently sparse, the number of variables in the dataset.
We then illustrate the main functions and algorithms on both real and simulated datasets.
arXiv Detail & Related papers (2022-01-28T09:30:32Z) - Hybrid Bayesian network discovery with latent variables by scoring
multiple interventions [5.994412766684843]
We present the hybrid mFGS-BS (majority rule and Fast Greedy equivalence Search with Bayesian Scoring) algorithm for structure learning from discrete data.
The algorithm assumes causal insufficiency in the presence of latent variables and produces a Partial Ancestral Graph (PAG)
Experimental results show that mFGS-BS improves structure learning accuracy relative to the state-of-the-art and it is computationally efficient.
arXiv Detail & Related papers (2021-12-20T14:54:41Z) - Evaluating State-of-the-Art Classification Models Against Bayes
Optimality [106.50867011164584]
We show that we can compute the exact Bayes error of generative models learned using normalizing flows.
We use our approach to conduct a thorough investigation of state-of-the-art classification models.
arXiv Detail & Related papers (2021-06-07T06:21:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.