Language Models as Fact Checkers?
- URL: http://arxiv.org/abs/2006.04102v2
- Date: Fri, 24 Jul 2020 07:15:37 GMT
- Title: Language Models as Fact Checkers?
- Authors: Nayeon Lee, Belinda Z. Li, Sinong Wang, Wen-tau Yih, Hao Ma, Madian
Khabsa
- Abstract summary: We create an effective end-to-end fact checker using a solely a language model.
We show that our zero-shot LM approach outperforms a random baseline on the standard FEVER task.
- Score: 39.29607585655352
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent work has suggested that language models (LMs) store both common-sense
and factual knowledge learned from pre-training data. In this paper, we
leverage this implicit knowledge to create an effective end-to-end fact checker
using a solely a language model, without any external knowledge or explicit
retrieval components. While previous work on extracting knowledge from LMs have
focused on the task of open-domain question answering, to the best of our
knowledge, this is the first work to examine the use of language models as fact
checkers. In a closed-book setting, we show that our zero-shot LM approach
outperforms a random baseline on the standard FEVER task, and that our
fine-tuned LM compares favorably with standard baselines. Though we do not
ultimately outperform methods which use explicit knowledge bases, we believe
our exploration shows that this method is viable and has much room for
exploration.
Related papers
- FactLLaMA: Optimizing Instruction-Following Language Models with
External Knowledge for Automated Fact-Checking [10.046323978189847]
We propose combining the power of instruction-following language models with external evidence retrieval to enhance fact-checking performance.
Our approach involves leveraging search engines to retrieve relevant evidence for a given input claim.
Then, we instruct-tune an open-sourced language model, called LLaMA, using this evidence, enabling it to predict the veracity of the input claim more accurately.
arXiv Detail & Related papers (2023-09-01T04:14:39Z) - Self-Checker: Plug-and-Play Modules for Fact-Checking with Large Language Models [75.75038268227554]
Self-Checker is a framework comprising a set of plug-and-play modules that facilitate fact-checking.
This framework provides a fast and efficient way to construct fact-checking systems in low-resource environments.
arXiv Detail & Related papers (2023-05-24T01:46:07Z) - Knowledge Rumination for Pre-trained Language Models [77.55888291165462]
We propose a new paradigm dubbed Knowledge Rumination to help the pre-trained language model utilize related latent knowledge without retrieving it from the external corpus.
We apply the proposed knowledge rumination to various language models, including RoBERTa, DeBERTa, and GPT-3.
arXiv Detail & Related papers (2023-05-15T15:47:09Z) - Can LMs Learn New Entities from Descriptions? Challenges in Propagating
Injected Knowledge [72.63368052592004]
We study LMs' abilities to make inferences based on injected facts (or propagate those facts)
We find that existing methods for updating knowledge show little propagation of injected knowledge.
Yet, prepending entity definitions in an LM's context improves performance across all settings.
arXiv Detail & Related papers (2023-05-02T17:59:46Z) - Augmented Language Models: a Survey [55.965967655575454]
This survey reviews works in which language models (LMs) are augmented with reasoning skills and the ability to use tools.
We refer to them as Augmented Language Models (ALMs)
The missing token objective allows ALMs to learn to reason, use tools, and even act, while still performing standard natural language tasks.
arXiv Detail & Related papers (2023-02-15T18:25:52Z) - LMPriors: Pre-Trained Language Models as Task-Specific Priors [78.97143833642971]
We develop principled techniques for augmenting our models with suitable priors.
This is to encourage them to learn in ways that are compatible with our understanding of the world.
We draw inspiration from the recent successes of large-scale language models (LMs) to construct task-specific priors distilled from the rich knowledge of LMs.
arXiv Detail & Related papers (2022-10-22T19:09:18Z) - Probing Commonsense Knowledge in Pre-trained Language Models with
Sense-level Precision and Expanded Vocabulary [2.588973722689844]
We present a method for enriching Language Models with a grounded sense inventory (i.e., WordNet) available at the vocabulary level, without further training.
We find that LMs can learn non-trivial commonsense knowledge from self-supervision, covering numerous relations, and more effectively than comparable similarity-based approaches.
arXiv Detail & Related papers (2022-10-12T16:26:59Z) - Learning Natural Language Generation from Scratch [25.984828046001013]
This paper introduces TRUncated ReinForcement Learning for Language (TrufLL)
It is an original ap-proach to train conditional language models from scratch by only using reinforcement learning (RL)
arXiv Detail & Related papers (2021-09-20T08:46:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.