Information fusion between knowledge and data in Bayesian network
structure learning
- URL: http://arxiv.org/abs/2102.00473v1
- Date: Sun, 31 Jan 2021 15:45:29 GMT
- Title: Information fusion between knowledge and data in Bayesian network
structure learning
- Authors: Anthony C. Constantinou, Zhigao Guo, Neville K. Kitson
- Abstract summary: This paper describes and evaluates a set of information fusion methods that have been implemented in the open-source Bayesys structure learning system.
The results are illustrated both with limited and big data, with application to three BN structure learning algorithms available in Bayesys.
- Score: 5.994412766684843
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Bayesian Networks (BNs) have become a powerful technology for reasoning under
uncertainty, particularly in areas that require causal assumptions that enable
us to simulate the effect of intervention. The graphical structure of these
models can be determined by causal knowledge, learnt from data, or a
combination of both. While it seems plausible that the best approach in
constructing a causal graph involves combining knowledge with machine learning,
this approach remains underused in practice. This paper describes and evaluates
a set of information fusion methods that have been implemented in the
open-source Bayesys structure learning system. The methods enable users to
specify pre-existing knowledge and rule-based information that can be obtained
from heterogeneous sources, to constrain or guide structure learning. Each
method is assessed in terms of structure learning impact, including graphical
accuracy, model fitting, complexity and runtime. The results are illustrated
both with limited and big data, with application to three BN structure learning
algorithms available in Bayesys, and reveal interesting inconsistencies about
their effectiveness where the results obtained from graphical measures often
contradict those obtained from model fitting measures. While the overall
results show that information fusion methods become less effective with big
data due to higher learning accuracy rendering knowledge less important, some
information fusion methods do perform better with big data. Lastly, amongst the
main conclusions is the observation that reduced search space obtained from
knowledge constraints does not imply reduced computational complexity, which
can happen when the constraints set up a tension between what the data indicate
and what the constraints are trying to enforce.
Related papers
- Learning to refine domain knowledge for biological network inference [2.209921757303168]
Perturbation experiments allow biologists to discover causal relationships between variables of interest.
The sparsity and high dimensionality of these data pose significant challenges for causal structure learning algorithms.
We propose an amortized algorithm for refining domain knowledge, based on data observations.
arXiv Detail & Related papers (2024-10-18T12:53:23Z) - Multi-modal Causal Structure Learning and Root Cause Analysis [67.67578590390907]
We propose Mulan, a unified multi-modal causal structure learning method for root cause localization.
We leverage a log-tailored language model to facilitate log representation learning, converting log sequences into time-series data.
We also introduce a novel key performance indicator-aware attention mechanism for assessing modality reliability and co-learning a final causal graph.
arXiv Detail & Related papers (2024-02-04T05:50:38Z) - Causal discovery using dynamically requested knowledge [7.904709685523615]
Causal Bayesian Networks (CBNs) are an important tool for reasoning under uncertainty in complex real-world systems.
We investigate a novel approach where the structure learning algorithm itself dynamically identifies and requests knowledge for relationships that the algorithm identifies as uncertain.
We show that it offers considerable gains in structural accuracy, which are generally larger than those offered by existing approaches for integrating knowledge.
arXiv Detail & Related papers (2023-10-17T11:21:23Z) - Recognizing Unseen Objects via Multimodal Intensive Knowledge Graph
Propagation [68.13453771001522]
We propose a multimodal intensive ZSL framework that matches regions of images with corresponding semantic embeddings.
We conduct extensive experiments and evaluate our model on large-scale real-world data.
arXiv Detail & Related papers (2023-06-14T13:07:48Z) - Schema-aware Reference as Prompt Improves Data-Efficient Knowledge Graph
Construction [57.854498238624366]
We propose a retrieval-augmented approach, which retrieves schema-aware Reference As Prompt (RAP) for data-efficient knowledge graph construction.
RAP can dynamically leverage schema and knowledge inherited from human-annotated and weak-supervised data as a prompt for each sample.
arXiv Detail & Related papers (2022-10-19T16:40:28Z) - A Meta-Reinforcement Learning Algorithm for Causal Discovery [3.4806267677524896]
Causal structures can enable models to go beyond pure correlation-based inference.
Finding causal structures from data poses a significant challenge both in computational effort and accuracy.
We develop a meta-reinforcement learning algorithm that performs causal discovery by learning to perform interventions.
arXiv Detail & Related papers (2022-07-18T09:26:07Z) - Distributionally Robust Semi-Supervised Learning Over Graphs [68.29280230284712]
Semi-supervised learning (SSL) over graph-structured data emerges in many network science applications.
To efficiently manage learning over graphs, variants of graph neural networks (GNNs) have been developed recently.
Despite their success in practice, most of existing methods are unable to handle graphs with uncertain nodal attributes.
Challenges also arise due to distributional uncertainties associated with data acquired by noisy measurements.
A distributionally robust learning framework is developed, where the objective is to train models that exhibit quantifiable robustness against perturbations.
arXiv Detail & Related papers (2021-10-20T14:23:54Z) - Learning Neural Causal Models with Active Interventions [83.44636110899742]
We introduce an active intervention-targeting mechanism which enables a quick identification of the underlying causal structure of the data-generating process.
Our method significantly reduces the required number of interactions compared with random intervention targeting.
We demonstrate superior performance on multiple benchmarks from simulated to real-world data.
arXiv Detail & Related papers (2021-09-06T13:10:37Z) - Incorporating Causal Graphical Prior Knowledge into Predictive Modeling
via Simple Data Augmentation [92.96204497841032]
Causal graphs (CGs) are compact representations of the knowledge of the data generating processes behind the data distributions.
We propose a model-agnostic data augmentation method that allows us to exploit the prior knowledge of the conditional independence (CI) relations.
We experimentally show that the proposed method is effective in improving the prediction accuracy, especially in the small-data regime.
arXiv Detail & Related papers (2021-02-27T06:13:59Z) - How do some Bayesian Network machine learned graphs compare to causal
knowledge? [6.5745172279769255]
The graph of a Bayesian Network (BN) can be machine learned, determined by causal knowledge, or a combination of both.
This paper focuses on purely machine learned and purely knowledge-based BNs.
It investigates their differences in terms of graphical structure and how well the implied statistical models explain the data.
arXiv Detail & Related papers (2021-01-25T22:29:54Z) - Causal Discovery from Incomplete Data: A Deep Learning Approach [21.289342482087267]
Imputated Causal Learning is proposed to perform iterative missing data imputation and causal structure discovery.
We show that ICL can outperform state-of-the-art methods under different missing data mechanisms.
arXiv Detail & Related papers (2020-01-15T14:28:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.