Testing Sparsity Assumptions in Bayesian Networks
- URL: http://arxiv.org/abs/2307.06406v1
- Date: Wed, 12 Jul 2023 18:46:22 GMT
- Title: Testing Sparsity Assumptions in Bayesian Networks
- Authors: Luke Duttweiler, Sally W. Thurston, and Anthony Almudevar
- Abstract summary: This paper provides the properties of, and a debiasing procedure for, a hypothesis test that may be used to determine if the BN has max in-degree greater than 1.
A linear BN structure discovery workflow is suggested in which the investigator uses this hypothesis test to aid in selecting an appropriate structure discovery algorithm.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Bayesian network (BN) structure discovery algorithms typically either make
assumptions about the sparsity of the true underlying network, or are limited
by computational constraints to networks with a small number of variables.
While these sparsity assumptions can take various forms, frequently the
assumptions focus on an upper bound for the maximum in-degree of the underlying
graph $\nabla_G$. Theorem 2 in Duttweiler et. al. (2023) demonstrates that the
largest eigenvalue of the normalized inverse covariance matrix ($\Omega$) of a
linear BN is a lower bound for $\nabla_G$. Building on this result, this paper
provides the asymptotic properties of, and a debiasing procedure for, the
sample eigenvalues of $\Omega$, leading to a hypothesis test that may be used
to determine if the BN has max in-degree greater than 1. A linear BN structure
discovery workflow is suggested in which the investigator uses this hypothesis
test to aid in selecting an appropriate structure discovery algorithm. The
hypothesis test performance is evaluated through simulations and the workflow
is demonstrated on data from a human psoriasis study.
Related papers
- Minimax Hypothesis Testing for the Bradley-Terry-Luce Model [6.5990719141691825]
The Bradley-Terry-Luce (BTL) model is one of the most widely used models for ranking a collection of items or agents.
We propose a hypothesis test that determines whether a given pairwise comparison dataset, with $k$ comparisons per pair of agents, originates from an underlying BTL model.
arXiv Detail & Related papers (2024-10-10T20:28:05Z) - Bayesian Approach to Linear Bayesian Networks [3.8711489380602804]
The proposed approach iteratively estimates each element of the topological ordering from backward and its parent using the inverse of a partial covariance matrix.
The proposed method is demonstrated to outperform state-of-the-art frequentist approaches, such as the BHLSM, LISTEN, and TD algorithms in synthetic data.
arXiv Detail & Related papers (2023-11-27T08:10:53Z) - Stable Nonconvex-Nonconcave Training via Linear Interpolation [51.668052890249726]
This paper presents a theoretical analysis of linearahead as a principled method for stabilizing (large-scale) neural network training.
We argue that instabilities in the optimization process are often caused by the nonmonotonicity of the loss landscape and show how linear can help by leveraging the theory of nonexpansive operators.
arXiv Detail & Related papers (2023-10-20T12:45:12Z) - Validation Diagnostics for SBI algorithms based on Normalizing Flows [55.41644538483948]
This work proposes easy to interpret validation diagnostics for multi-dimensional conditional (posterior) density estimators based on NF.
It also offers theoretical guarantees based on results of local consistency.
This work should help the design of better specified models or drive the development of novel SBI-algorithms.
arXiv Detail & Related papers (2022-11-17T15:48:06Z) - SIMPLE-RC: Group Network Inference with Non-Sharp Nulls and Weak Signals [8.948321043168455]
We propose a SIMPLE method with random coupling (SIMPLE-RC) for testing the non-sharp null hypothesis.
We construct our test as the maximum of the SIMPLE tests for subsampled node pairs from the group.
New theoretical developments are empowered by a second-order expansion of spiked eigenvectors under the $ell_infty$-norm.
arXiv Detail & Related papers (2022-10-31T20:36:24Z) - Robust Training and Verification of Implicit Neural Networks: A
Non-Euclidean Contractive Approach [64.23331120621118]
This paper proposes a theoretical and computational framework for training and robustness verification of implicit neural networks.
We introduce a related embedded network and show that the embedded network can be used to provide an $ell_infty$-norm box over-approximation of the reachable sets of the original network.
We apply our algorithms to train implicit neural networks on the MNIST dataset and compare the robustness of our models with the models trained via existing approaches in the literature.
arXiv Detail & Related papers (2022-08-08T03:13:24Z) - On the Effective Number of Linear Regions in Shallow Univariate ReLU
Networks: Convergence Guarantees and Implicit Bias [50.84569563188485]
We show that gradient flow converges in direction when labels are determined by the sign of a target network with $r$ neurons.
Our result may already hold for mild over- parameterization, where the width is $tildemathcalO(r)$ and independent of the sample size.
arXiv Detail & Related papers (2022-05-18T16:57:10Z) - Predicting Unreliable Predictions by Shattering a Neural Network [145.3823991041987]
Piecewise linear neural networks can be split into subfunctions.
Subfunctions have their own activation pattern, domain, and empirical error.
Empirical error for the full network can be written as an expectation over subfunctions.
arXiv Detail & Related papers (2021-06-15T18:34:41Z) - Structure Learning in Inverse Ising Problems Using $\ell_2$-Regularized
Linear Estimator [8.89493507314525]
We show that despite the model mismatch, one can perfectly identify the network structure using naive linear regression without regularization.
We propose a two-stage estimator: In the first stage, the ridge regression is used and the estimates are pruned by a relatively small threshold.
This estimator with the appropriate regularization coefficient and thresholds is shown to achieve the perfect identification of the network structure even in $0M/N1$.
arXiv Detail & Related papers (2020-08-19T09:11:33Z) - On the Convergence Rate of Projected Gradient Descent for a
Back-Projection based Objective [58.33065918353532]
We consider a back-projection based fidelity term as an alternative to the common least squares (LS)
We show that using the BP term, rather than the LS term, requires fewer iterations of optimization algorithms.
arXiv Detail & Related papers (2020-05-03T00:58:23Z) - A Neural Network Based on First Principles [13.554038901140949]
A Neural network is derived from first principles, assuming only that each layer begins with a linear dimension-reducing transformation.
The approach appeals to the principle of Maximum Entropy (MaxEnt) to find the posterior distribution of the input data of each layer.
arXiv Detail & Related papers (2020-02-18T10:16:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.