CausNet : Generational orderings based search for optimal Bayesian
networks via dynamic programming with parent set constraints
- URL: http://arxiv.org/abs/2207.08365v1
- Date: Mon, 18 Jul 2022 03:26:41 GMT
- Title: CausNet : Generational orderings based search for optimal Bayesian
networks via dynamic programming with parent set constraints
- Authors: Nand Sharma, Joshua Millstein
- Abstract summary: We implement a dynamic programming based algorithm with built-in dimensionality reduction and parent set identification.
This reduces the search space drastically and can be applied to large-dimensional data.
We demonstrate the efficacy of our algorithm on both synthetic and real data.
- Score: 5.7460673927330035
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Finding a globally optimal Bayesian Network using exhaustive search is a
problem with super-exponential complexity, which severely restricts the number
of variables that it can work for. We implement a dynamic programming based
algorithm with built-in dimensionality reduction and parent set identification.
This reduces the search space drastically and can be applied to
large-dimensional data. We use what we call generational orderings based search
for optimal networks, which is a novel way to efficiently search the space of
possible networks given the possible parent sets. The algorithm supports both
continuous and categorical data, and categorical as well as survival outcomes.
We demonstrate the efficacy of our algorithm on both synthetic and real data.
In simulations, our algorithm performs better than three state-of-art
algorithms that are currently used extensively. We then apply it to an Ovarian
Cancer gene expression dataset with 513 genes and a survival outcome. Our
algorithm is able to find an optimal network describing the disease pathway
consisting of 6 genes leading to the outcome node in a few minutes on a basic
computer. Our generational orderings based search for optimal networks, is both
efficient and highly scalable approach to finding optimal Bayesian Networks,
that can be applied to 1000s of variables. Using specifiable parameters -
correlation, FDR cutoffs, and in-degree - one can increase or decrease the
number of nodes and density of the networks. Availability of two scoring
option-BIC and Bge-and implementation of survival outcomes and mixed data types
makes our algorithm very suitable for many types of high dimensional biomedical
data to find disease pathways.
Related papers
- GDSG: Graph Diffusion-based Solution Generator for Optimization Problems in MEC Networks [109.17835015018532]
We present a Graph Diffusion-based Solution Generation (GDSG) method.
This approach is designed to work with suboptimal datasets while converging to the optimal solution large probably.
We build GDSG as a multi-task diffusion model utilizing a Graph Neural Network (GNN) to acquire the distribution of high-quality solutions.
arXiv Detail & Related papers (2024-12-11T11:13:43Z) - Exact discovery is polynomial for certain sparse causal Bayesian networks [1.5293427903448027]
We show how to use properties of Bayesian networks to prune the search space and lower the computational cost.
We also include new path-search and divide-and-conquer criteria.
Our approach then out-competes the state-of-the-art at lower densities.
arXiv Detail & Related papers (2024-06-21T09:41:30Z) - A Hybrid Genetic Algorithm for the min-max Multiple Traveling Salesman
Problem [0.9790236766474201]
This paper proposes a hybrid genetic algorithm for solving the Multiple Traveling Salesman Problem (mTSP)
A novel crossover operator is designed to combine similar tours from two parents and offers great diversity for the population.
Our algorithm outperforms all existing algorithms on average, with similar cutoff time thresholds, when tested against benchmark sets.
arXiv Detail & Related papers (2023-07-14T01:57:10Z) - BigBraveBN: algorithm of structural learning for bayesian networks with
a large number of nodes [0.0]
The article presents a BigBraveBN algorithm for learning large Bayesian Networks with a high number of nodes (over 100)
The algorithm utilizes the Brave coefficient that measures the mutual occurrence of instances in several groups.
In the experimental part of the article, we compare the performance of BigBraveBN to other existing solutions on multiple data sets both discrete and continuous.
arXiv Detail & Related papers (2022-08-22T13:43:57Z) - A Metaheuristic Algorithm for Large Maximum Weight Independent Set
Problems [58.348679046591265]
Given a node-weighted graph, find a set of independent (mutually nonadjacent) nodes whose node-weight sum is maximum.
Some of the graphs airsing in this application are large, having hundreds of thousands of nodes and hundreds of millions of edges.
We develop a new local search algorithm, which is a metaheuristic in the greedy randomized adaptive search framework.
arXiv Detail & Related papers (2022-03-28T21:34:16Z) - A Sparse Structure Learning Algorithm for Bayesian Network
Identification from Discrete High-Dimensional Data [0.40611352512781856]
This paper addresses the problem of learning a sparse structure Bayesian network from high-dimensional discrete data.
We propose a score function that satisfies the sparsity and the DAG property simultaneously.
Specifically, we use a variance reducing method in our optimization algorithm to make the algorithm work efficiently in high-dimensional data.
arXiv Detail & Related papers (2021-08-21T12:21:01Z) - Unsupervised Domain-adaptive Hash for Networks [81.49184987430333]
Domain-adaptive hash learning has enjoyed considerable success in the computer vision community.
We develop an unsupervised domain-adaptive hash learning method for networks, dubbed UDAH.
arXiv Detail & Related papers (2021-08-20T12:09:38Z) - Waypoint Planning Networks [66.72790309889432]
We propose a hybrid algorithm based on LSTMs with a local kernel - a classic algorithm such as A*, and a global kernel using a learned algorithm.
We compare WPN against A*, as well as related works including motion planning networks (MPNet) and value networks (VIN)
It is shown that WPN's search space is considerably less than A*, while being able to generate near optimal results.
arXiv Detail & Related papers (2021-05-01T18:02:01Z) - Towards Optimally Efficient Tree Search with Deep Learning [76.64632985696237]
This paper investigates the classical integer least-squares problem which estimates signals integer from linear models.
The problem is NP-hard and often arises in diverse applications such as signal processing, bioinformatics, communications and machine learning.
We propose a general hyper-accelerated tree search (HATS) algorithm by employing a deep neural network to estimate the optimal estimation for the underlying simplified memory-bounded A* algorithm.
arXiv Detail & Related papers (2021-01-07T08:00:02Z) - Online Model Selection for Reinforcement Learning with Function
Approximation [50.008542459050155]
We present a meta-algorithm that adapts to the optimal complexity with $tildeO(L5/6 T2/3)$ regret.
We also show that the meta-algorithm automatically admits significantly improved instance-dependent regret bounds.
arXiv Detail & Related papers (2020-11-19T10:00:54Z) - Genetic-algorithm-optimized neural networks for gravitational wave
classification [0.0]
We propose a new method for hyperparameter optimization based on genetic algorithms (GAs)
We show that the GA can discover high-quality architectures when the initial hyper parameter seed values are far from a good solution.
Using genetic algorithm optimization to refine an existing network should be especially useful if the problem context changes.
arXiv Detail & Related papers (2020-10-09T03:14:20Z) - A Systematic Characterization of Sampling Algorithms for Open-ended
Language Generation [71.31905141672529]
We study the widely adopted ancestral sampling algorithms for auto-regressive language models.
We identify three key properties that are shared among them: entropy reduction, order preservation, and slope preservation.
We find that the set of sampling algorithms that satisfies these properties performs on par with the existing sampling algorithms.
arXiv Detail & Related papers (2020-09-15T17:28:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.