Causal Discovery with Language Models as Imperfect Experts
- URL: http://arxiv.org/abs/2307.02390v1
- Date: Wed, 5 Jul 2023 16:01:38 GMT
- Title: Causal Discovery with Language Models as Imperfect Experts
- Authors: Stephanie Long, Alexandre Pich\'e, Valentina Zantedeschi, Tibor
Schuster, Alexandre Drouin
- Abstract summary: We consider how expert knowledge can be used to improve the data-driven identification of causal graphs.
We propose strategies for amending such expert knowledge based on consistency properties.
We report a case study, on real data, where a large language model is used as an imperfect expert.
- Score: 119.22928856942292
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Understanding the causal relationships that underlie a system is a
fundamental prerequisite to accurate decision-making. In this work, we explore
how expert knowledge can be used to improve the data-driven identification of
causal graphs, beyond Markov equivalence classes. In doing so, we consider a
setting where we can query an expert about the orientation of causal
relationships between variables, but where the expert may provide erroneous
information. We propose strategies for amending such expert knowledge based on
consistency properties, e.g., acyclicity and conditional independencies in the
equivalence class. We then report a case study, on real data, where a large
language model is used as an imperfect expert.
Related papers
- Does Knowledge Localization Hold True? Surprising Differences Between Entity and Relation Perspectives in Language Models [20.157061521694096]
This study investigates the differences between entity and relational knowledge through knowledge editing.
To further elucidate the differences between entity and relational knowledge, we employ causal analysis to investigate how relational knowledge is stored in pre-trained models.
This insight highlights the multifaceted nature of knowledge storage in language models, underscoring the complexity of manipulating specific types of knowledge within these models.
arXiv Detail & Related papers (2024-09-01T05:09:11Z) - Unsupervised Pairwise Causal Discovery on Heterogeneous Data using Mutual Information Measures [49.1574468325115]
Causal Discovery is a technique that tackles the challenge by analyzing the statistical properties of the constituent variables.
We question the current (possibly misleading) baseline results on the basis that they were obtained through supervised learning.
In consequence, we approach this problem in an unsupervised way, using robust Mutual Information measures.
arXiv Detail & Related papers (2024-08-01T09:11:08Z) - Large Language Models are Limited in Out-of-Context Knowledge Reasoning [65.72847298578071]
Large Language Models (LLMs) possess extensive knowledge and strong capabilities in performing in-context reasoning.
This paper focuses on a significant aspect of out-of-context reasoning: Out-of-Context Knowledge Reasoning (OCKR), which is to combine multiple knowledge to infer new knowledge.
arXiv Detail & Related papers (2024-06-11T15:58:59Z) - Detection and Evaluation of bias-inducing Features in Machine learning [14.045499740240823]
In the context of machine learning (ML), one can use cause-to-effect analysis to understand the reason for the biased behavior of the system.
We propose an approach for systematically identifying all bias-inducing features of a model to help support the decision-making of domain experts.
arXiv Detail & Related papers (2023-10-19T15:01:16Z) - Probabilistic Dataset Reconstruction from Interpretable Models [8.31111379034875]
We show that optimal interpretable models are often more compact and leak less information regarding their training data than greedily-built ones.
Our results suggest that optimal interpretable models are often more compact and leak less information regarding their training data than greedily-built ones.
arXiv Detail & Related papers (2023-08-29T08:10:09Z) - Leveraging Skill-to-Skill Supervision for Knowledge Tracing [13.753990664747265]
Knowledge tracing plays a pivotal role in intelligent tutoring systems.
Recent advances in knowledge tracing models have enabled better exploitation of problem solving history.
Knowledge tracing algorithms that incorporate knowledge directly are important to settings with limited data or cold starts.
arXiv Detail & Related papers (2023-06-12T03:23:22Z) - A Conceptual Model for End-to-End Causal Discovery in Knowledge Tracing [8.049552839071918]
We take a preliminary step towards solving the problem of causal discovery in knowledge tracing.
Our solution placed among the top entries in Task 3 of the NeurIPS 2022 Challenge on Causal Insights for Learning Paths in Education.
arXiv Detail & Related papers (2023-05-11T21:20:29Z) - Competency Problems: On Finding and Removing Artifacts in Language Data [50.09608320112584]
We argue that for complex language understanding tasks, all simple feature correlations are spurious.
We theoretically analyze the difficulty of creating data for competency problems when human bias is taken into account.
arXiv Detail & Related papers (2021-04-17T21:34:10Z) - Leveraging Expert Consistency to Improve Algorithmic Decision Support [62.61153549123407]
We explore the use of historical expert decisions as a rich source of information that can be combined with observed outcomes to narrow the construct gap.
We propose an influence function-based methodology to estimate expert consistency indirectly when each case in the data is assessed by a single expert.
Our empirical evaluation, using simulations in a clinical setting and real-world data from the child welfare domain, indicates that the proposed approach successfully narrows the construct gap.
arXiv Detail & Related papers (2021-01-24T05:40:29Z) - Knowledge-driven Data Construction for Zero-shot Evaluation in
Commonsense Question Answering [80.60605604261416]
We propose a novel neuro-symbolic framework for zero-shot question answering across commonsense tasks.
We vary the set of language models, training regimes, knowledge sources, and data generation strategies, and measure their impact across tasks.
We show that, while an individual knowledge graph is better suited for specific tasks, a global knowledge graph brings consistent gains across different tasks.
arXiv Detail & Related papers (2020-11-07T22:52:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.