FSPN: A New Class of Probabilistic Graphical Model
- URL: http://arxiv.org/abs/2011.09020v2
- Date: Fri, 20 Nov 2020 08:22:09 GMT
- Title: FSPN: A New Class of Probabilistic Graphical Model
- Authors: Ziniu Wu, Rong Zhu, Andreas Pfadler, Yuxing Han, Jiangneng Li,
Zhengping Qian, Kai Zeng, Jingren Zhou
- Abstract summary: We introduce factorize sum split product networks (FSPNs), a new class of probabilistic graphical models (PGMs)
FSPNs are designed to overcome the drawbacks of existing PGMs in terms of estimation accuracy and inference efficiency.
We present efficient probability inference and structure learning algorithms for FSPNs, along with a theoretical analysis and extensive evaluation evidence.
- Score: 37.80683263600885
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We introduce factorize sum split product networks (FSPNs), a new class of
probabilistic graphical models (PGMs). FSPNs are designed to overcome the
drawbacks of existing PGMs in terms of estimation accuracy and inference
efficiency. Specifically, Bayesian networks (BNs) have low inference speed and
performance of tree structured sum product networks(SPNs) significantly
degrades in presence of highly correlated variables. FSPNs absorb their
advantages by adaptively modeling the joint distribution of variables according
to their dependence degree, so that one can simultaneously attain the two
desirable goals: high estimation accuracy and fast inference speed. We present
efficient probability inference and structure learning algorithms for FSPNs,
along with a theoretical analysis and extensive evaluation evidence. Our
experimental results on synthetic and benchmark datasets indicate the
superiority of FSPN over other PGMs.
Related papers
- Improving Tree Probability Estimation with Stochastic Optimization and Variance Reduction [11.417249588622926]
Subsplit Bayesian networks (SBNs) provide a powerful probabilistic graphical model for tree probability estimation.
The expectation (EM) method currently used for learning SBN parameters does not scale up to large data sets.
We introduce several computationally efficient methods for training SBNs and show that variance reduction could be the key for better performance.
arXiv Detail & Related papers (2024-09-09T02:22:52Z) - Enhancing Fast Feed Forward Networks with Load Balancing and a Master Leaf Node [49.08777822540483]
Fast feedforward networks (FFFs) exploit the observation that different regions of the input space activate distinct subsets of neurons in wide networks.
We propose the incorporation of load balancing and Master Leaf techniques into the FFF architecture to improve performance and simplify the training process.
arXiv Detail & Related papers (2024-05-27T05:06:24Z) - Chasing Fairness in Graphs: A GNN Architecture Perspective [73.43111851492593]
We propose textsfFair textsfMessage textsfPassing (FMP) designed within a unified optimization framework for graph neural networks (GNNs)
In FMP, the aggregation is first adopted to utilize neighbors' information and then the bias mitigation step explicitly pushes demographic group node presentation centers together.
Experiments on node classification tasks demonstrate that the proposed FMP outperforms several baselines in terms of fairness and accuracy on three real-world datasets.
arXiv Detail & Related papers (2023-12-19T18:00:15Z) - Layer-wise Feedback Propagation [53.00944147633484]
We present Layer-wise Feedback Propagation (LFP), a novel training approach for neural-network-like predictors.
LFP assigns rewards to individual connections based on their respective contributions to solving a given task.
We demonstrate its effectiveness in achieving comparable performance to gradient descent on various models and datasets.
arXiv Detail & Related papers (2023-08-23T10:48:28Z) - A New PHO-rmula for Improved Performance of Semi-Structured Networks [0.0]
We show that techniques to properly identify the contributions of the different model components in SSNs lead to suboptimal network estimation.
We propose a non-invasive post-hocization (PHO) that guarantees identifiability of model components and provides better estimation and prediction quality.
Our theoretical findings are supported by numerical experiments, a benchmark comparison as well as a real-world application to COVID-19 infections.
arXiv Detail & Related papers (2023-06-01T10:23:28Z) - Tractable Probabilistic Graph Representation Learning with Graph-Induced
Sum-Product Networks [25.132159381873656]
We introduce Graph-Induced Sum-Product Networks (GSPNs), a new probabilistic framework for graph representation learning.
We show the model's competitiveness on scarce supervision scenarios, under missing data, and for graph classification in comparison to popular neural models.
arXiv Detail & Related papers (2023-05-17T20:02:08Z) - Performance and Interpretability Comparisons of Supervised Machine
Learning Algorithms: An Empirical Study [3.7881729884531805]
The paper is organized in a findings-based manner, with each section providing general conclusions.
Overall, XGB and FFNNs were competitive, with FFNNs showing better performance in smooth models.
RF did not perform well in general, confirming the findings in the literature.
arXiv Detail & Related papers (2022-04-27T12:04:33Z) - Hybrid Bayesian network discovery with latent variables by scoring
multiple interventions [5.994412766684843]
We present the hybrid mFGS-BS (majority rule and Fast Greedy equivalence Search with Bayesian Scoring) algorithm for structure learning from discrete data.
The algorithm assumes causal insufficiency in the presence of latent variables and produces a Partial Ancestral Graph (PAG)
Experimental results show that mFGS-BS improves structure learning accuracy relative to the state-of-the-art and it is computationally efficient.
arXiv Detail & Related papers (2021-12-20T14:54:41Z) - Accuracy on the Line: On the Strong Correlation Between
Out-of-Distribution and In-Distribution Generalization [89.73665256847858]
We show that out-of-distribution performance is strongly correlated with in-distribution performance for a wide range of models and distribution shifts.
Specifically, we demonstrate strong correlations between in-distribution and out-of-distribution performance on variants of CIFAR-10 & ImageNet.
We also investigate cases where the correlation is weaker, for instance some synthetic distribution shifts from CIFAR-10-C and the tissue classification dataset Camelyon17-WILDS.
arXiv Detail & Related papers (2021-07-09T19:48:23Z) - Probabilistic Gradient Boosting Machines for Large-Scale Probabilistic
Regression [51.770998056563094]
Probabilistic Gradient Boosting Machines (PGBM) is a method to create probabilistic predictions with a single ensemble of decision trees.
We empirically demonstrate the advantages of PGBM compared to existing state-of-the-art methods.
arXiv Detail & Related papers (2021-06-03T08:32:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.