BeliefBank: Adding Memory to a Pre-Trained Language Model for a
Systematic Notion of Belief
- URL: http://arxiv.org/abs/2109.14723v1
- Date: Wed, 29 Sep 2021 21:04:27 GMT
- Title: BeliefBank: Adding Memory to a Pre-Trained Language Model for a
Systematic Notion of Belief
- Authors: Nora Kassner, Oyvind Tafjord, Hinrich Sch\"utze, Peter Clark
- Abstract summary: It can be hard to identify what the model actually "believes" about the world, making it susceptible to inconsistent behavior and simple errors.
Our approach is to embed a PTLM in a broader system that includes an evolving, symbolic memory of beliefs.
We show that, in a controlled experimental setting, these two mechanisms result in more consistent beliefs in the overall system.
- Score: 20.60798513220516
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Although pretrained language models (PTLMs) contain significant amounts of
world knowledge, they can still produce inconsistent answers to questions when
probed, even after specialized training. As a result, it can be hard to
identify what the model actually "believes" about the world, making it
susceptible to inconsistent behavior and simple errors. Our goal is to reduce
these problems. Our approach is to embed a PTLM in a broader system that also
includes an evolving, symbolic memory of beliefs -- a BeliefBank -- that
records but then may modify the raw PTLM answers. We describe two mechanisms to
improve belief consistency in the overall system. First, a reasoning component
-- a weighted MaxSAT solver -- revises beliefs that significantly clash with
others. Second, a feedback component issues future queries to the PTLM using
known beliefs as context. We show that, in a controlled experimental setting,
these two mechanisms result in more consistent beliefs in the overall system,
improving both the accuracy and consistency of its answers over time. This is
significant as it is a first step towards PTLM-based architectures with a
systematic notion of belief, enabling them to construct a more coherent picture
of the world, and improve over time without model retraining.
Related papers
- Belief Revision: The Adaptability of Large Language Models Reasoning [63.0281286287648]
We introduce Belief-R, a new dataset designed to test LMs' belief revision ability when presented with new evidence.
Inspired by how humans suppress prior inferences, this task assesses LMs within the newly proposed delta reasoning framework.
We evaluate $sim$30 LMs across diverse prompting strategies and found that LMs generally struggle to appropriately revise their beliefs in response to new information.
arXiv Detail & Related papers (2024-06-28T09:09:36Z) - Can Large Language Models Always Solve Easy Problems if They Can Solve Harder Ones? [65.43882564649721]
Large language models (LLMs) have demonstrated impressive capabilities, but still suffer from inconsistency issues.
We develop the ConsisEval benchmark, where each entry comprises a pair of questions with a strict order of difficulty.
We analyze the potential for improvement in consistency by relative consistency score.
arXiv Detail & Related papers (2024-06-18T17:25:47Z) - Distilling Reasoning Ability from Large Language Models with Adaptive Thinking [54.047761094420174]
Chain of thought finetuning (cot-finetuning) aims to endow small language models (SLM) with reasoning ability to improve their performance towards specific tasks.
Most existing cot-finetuning methods adopt a pre-thinking mechanism, allowing the SLM to generate a rationale before providing an answer.
This mechanism enables SLM to analyze and think about complex questions, but it also makes answer correctness highly sensitive to minor errors in rationale.
We propose a robust post-thinking mechanism to generate answers before rationale.
arXiv Detail & Related papers (2024-04-14T07:19:27Z) - Improving Language Models Meaning Understanding and Consistency by
Learning Conceptual Roles from Dictionary [65.268245109828]
Non-human-like behaviour of contemporary pre-trained language models (PLMs) is a leading cause undermining their trustworthiness.
A striking phenomenon is the generation of inconsistent predictions, which produces contradictory results.
We propose a practical approach that alleviates the inconsistent behaviour issue by improving PLM awareness.
arXiv Detail & Related papers (2023-10-24T06:15:15Z) - Language Models with Rationality [57.37201135072838]
Large language models (LLMs) are proficient at question-answering (QA)
It is not always clear how (or even if) an answer follows from their latent "beliefs"
arXiv Detail & Related papers (2023-05-23T17:04:25Z) - Entailer: Answering Questions with Faithful and Truthful Chains of
Reasoning [26.715242799194908]
We show how a question-answering system can show how its answers are implied by its own internal beliefs via a systematic chain of reasoning.
Our approach is to combine a trained backward-chaining model, capable of generating a set of premises entailing an answer hypothesis, with a verifier that checks that the model itself believes those premises.
To our knowledge, this is the first system to generate multistep chains that are both faithful (the answer follows from the reasoning) and truthful (the chain reflects the system's own internal beliefs)
arXiv Detail & Related papers (2022-10-21T19:51:56Z) - Towards Teachable Reasoning Systems [29.59387051046722]
We develop a teachable reasoning system for question-answering (QA)
Our approach is three-fold: First, generated chains of reasoning show how answers are implied by the system's own internal beliefs.
Second, users can interact with the explanations to identify erroneous model beliefs and provide corrections.
Third, we augment the model with a dynamic memory of such corrections.
arXiv Detail & Related papers (2022-04-27T17:15:07Z) - Do Language Models Have Beliefs? Methods for Detecting, Updating, and
Visualizing Model Beliefs [76.6325846350907]
Dennett (1995) famously argues that even thermostats have beliefs, on the view that a belief is simply an informational state decoupled from any motivational state.
In this paper, we discuss approaches to detecting when models have beliefs about the world, and we improve on methods for updating model beliefs to be more truthful.
arXiv Detail & Related papers (2021-11-26T18:33:59Z) - Enriching a Model's Notion of Belief using a Persistent Memory [20.60798513220516]
Pretrained language models (PTLMs) can produce inconsistent answers to questions when probed.
It can be hard to identify what the model actually "believes" about the world.
Our goal is to reduce this problem, so systems are more globally consistent and accurate in their answers.
arXiv Detail & Related papers (2021-04-16T23:09:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.