Characteristic Circuits
- URL: http://arxiv.org/abs/2312.07790v1
- Date: Tue, 12 Dec 2023 23:15:07 GMT
- Title: Characteristic Circuits
- Authors: Zhongjie Yu, Martin Trapp, Kristian Kersting
- Abstract summary: Probabilistic circuits (PCs) compose simple, tractable distributions into a high-dimensional probability distribution.
We introduce characteristic circuits (CCs) providing a unified formalization of distributions over heterogeneous data in the spectral domain.
We show that CCs outperform state-of-the-art density estimators for heterogeneous data domains on common benchmark data sets.
- Score: 26.223089423713486
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In many real-world scenarios, it is crucial to be able to reliably and
efficiently reason under uncertainty while capturing complex relationships in
data. Probabilistic circuits (PCs), a prominent family of tractable
probabilistic models, offer a remedy to this challenge by composing simple,
tractable distributions into a high-dimensional probability distribution.
However, learning PCs on heterogeneous data is challenging and densities of
some parametric distributions are not available in closed form, limiting their
potential use. We introduce characteristic circuits (CCs), a family of
tractable probabilistic models providing a unified formalization of
distributions over heterogeneous data in the spectral domain. The one-to-one
relationship between characteristic functions and probability measures enables
us to learn high-dimensional distributions on heterogeneous data domains and
facilitates efficient probabilistic inference even when no closed-form density
function is available. We show that the structure and parameters of CCs can be
learned efficiently from the data and find that CCs outperform state-of-the-art
density estimators for heterogeneous data domains on common benchmark data
sets.
Related papers
- Probabilistically Plausible Counterfactual Explanations with Normalizing Flows [2.675793767640172]
We present PPCEF, a novel method for generating probabilistically plausible counterfactual explanations.
Our method enforces plausibility by directly optimizing the explicit density function without assuming a particular family of parametrized distributions.
PPCEF is a powerful tool for interpreting machine learning models and for improving fairness, accountability, and trust in AI systems.
arXiv Detail & Related papers (2024-05-27T20:24:03Z) - Federated Causal Discovery from Heterogeneous Data [70.31070224690399]
We propose a novel FCD method attempting to accommodate arbitrary causal models and heterogeneous data.
These approaches involve constructing summary statistics as a proxy of the raw data to protect data privacy.
We conduct extensive experiments on synthetic and real datasets to show the efficacy of our method.
arXiv Detail & Related papers (2024-02-20T18:53:53Z) - Learning to Bound Counterfactual Inference in Structural Causal Models
from Observational and Randomised Data [64.96984404868411]
We derive a likelihood characterisation for the overall data that leads us to extend a previous EM-based algorithm.
The new algorithm learns to approximate the (unidentifiability) region of model parameters from such mixed data sources.
It delivers interval approximations to counterfactual results, which collapse to points in the identifiable case.
arXiv Detail & Related papers (2022-12-06T12:42:11Z) - Learning Multivariate CDFs and Copulas using Tensor Factorization [39.24470798045442]
Learning the multivariate distribution of data is a core challenge in statistics and machine learning.
In this work, we aim to learn multivariate cumulative distribution functions (CDFs), as they can handle mixed random variables.
We show that any grid sampled version of a joint CDF of mixed random variables admits a universal representation as a naive Bayes model.
We demonstrate the superior performance of the proposed model in several synthetic and real datasets and applications including regression, sampling and data imputation.
arXiv Detail & Related papers (2022-10-13T16:18:46Z) - Marginalization in Bayesian Networks: Integrating Exact and Approximate
Inference [0.0]
Missing data and hidden variables require calculating the marginal probability distribution of a subset of the variables.
We develop a divide-and-conquer approach using the graphical properties of Bayesian networks.
We present an efficient and scalable algorithm for estimating the marginal probability distribution for categorical variables.
arXiv Detail & Related papers (2021-12-16T21:49:52Z) - Probabilistic Generating Circuits [50.98473654244851]
We propose probabilistic generating circuits (PGCs) for their efficient representation.
PGCs are not just a theoretical framework that unifies vastly different existing models, but also show huge potential in modeling realistic data.
We exhibit a simple class of PGCs that are not trivially subsumed by simple combinations of PCs and DPPs, and obtain competitive performance on a suite of density estimation benchmarks.
arXiv Detail & Related papers (2021-02-19T07:06:53Z) - General stochastic separation theorems with optimal bounds [68.8204255655161]
Phenomenon of separability was revealed and used in machine learning to correct errors of Artificial Intelligence (AI) systems and analyze AI instabilities.
Errors or clusters of errors can be separated from the rest of the data.
The ability to correct an AI system also opens up the possibility of an attack on it, and the high dimensionality induces vulnerabilities caused by the same separability.
arXiv Detail & Related papers (2020-10-11T13:12:41Z) - Bayesian Sparse Factor Analysis with Kernelized Observations [67.60224656603823]
Multi-view problems can be faced with latent variable models.
High-dimensionality and non-linear issues are traditionally handled by kernel methods.
We propose merging both approaches into single model.
arXiv Detail & Related papers (2020-06-01T14:25:38Z) - Distribution Approximation and Statistical Estimation Guarantees of
Generative Adversarial Networks [82.61546580149427]
Generative Adversarial Networks (GANs) have achieved a great success in unsupervised learning.
This paper provides approximation and statistical guarantees of GANs for the estimation of data distributions with densities in a H"older space.
arXiv Detail & Related papers (2020-02-10T16:47:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.