SciFact-Open: Towards open-domain scientific claim verification
- URL: http://arxiv.org/abs/2210.13777v1
- Date: Tue, 25 Oct 2022 05:45:00 GMT
- Title: SciFact-Open: Towards open-domain scientific claim verification
- Authors: David Wadden, Kyle Lo, Bailey Kuehl, Arman Cohan, Iz Beltagy, Lucy Lu
Wang, Hannaneh Hajishirzi
- Abstract summary: We present SciFact-Open, a new test collection designed to evaluate the performance of scientific claim verification systems.
We collect evidence for scientific claims by pooling and annotating the top predictions of four state-of-the-art scientific claim verification models.
We find that systems developed on smaller corpora struggle to generalize to SciFact-Open, exhibiting performance drops of at least 15 F1.
- Score: 61.288725621156864
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: While research on scientific claim verification has led to the development of
powerful systems that appear to approach human performance, these approaches
have yet to be tested in a realistic setting against large corpora of
scientific literature. Moving to this open-domain evaluation setting, however,
poses unique challenges; in particular, it is infeasible to exhaustively
annotate all evidence documents. In this work, we present SciFact-Open, a new
test collection designed to evaluate the performance of scientific claim
verification systems on a corpus of 500K research abstracts. Drawing upon
pooling techniques from information retrieval, we collect evidence for
scientific claims by pooling and annotating the top predictions of four
state-of-the-art scientific claim verification models. We find that systems
developed on smaller corpora struggle to generalize to SciFact-Open, exhibiting
performance drops of at least 15 F1. In addition, analysis of the evidence in
SciFact-Open reveals interesting phenomena likely to appear when claim
verification systems are deployed in practice, e.g., cases where the evidence
supports only a special case of the claim. Our dataset is available at
https://github.com/dwadden/scifact-open.
Related papers
- Comparing Knowledge Sources for Open-Domain Scientific Claim
Verification [6.726255259929497]
We show that PubMed works better with specialized biomedical claims, while Wikipedia is more suited for everyday health concerns.
We discuss the results, outline frequent retrieval patterns and challenges, and provide promising future directions.
arXiv Detail & Related papers (2024-02-05T09:57:15Z) - Large Language Models for Automated Open-domain Scientific Hypotheses Discovery [50.40483334131271]
This work proposes the first dataset for social science academic hypotheses discovery.
Unlike previous settings, the new dataset requires (1) using open-domain data (raw web corpus) as observations; and (2) proposing hypotheses even new to humanity.
A multi- module framework is developed for the task, including three different feedback mechanisms to boost performance.
arXiv Detail & Related papers (2023-09-06T05:19:41Z) - SCITAB: A Challenging Benchmark for Compositional Reasoning and Claim
Verification on Scientific Tables [68.76415918462418]
We present SCITAB, a challenging evaluation dataset consisting of 1.2K expert-verified scientific claims.
Through extensive evaluations, we demonstrate that SCITAB poses a significant challenge to state-of-the-art models.
Our analysis uncovers several unique challenges posed by SCITAB, including table grounding, claim ambiguity, and compositional reasoning.
arXiv Detail & Related papers (2023-05-22T16:13:50Z) - Modeling Information Change in Science Communication with Semantically
Matched Paraphrases [50.67030449927206]
SPICED is the first paraphrase dataset of scientific findings annotated for degree of information change.
SPICED contains 6,000 scientific finding pairs extracted from news stories, social media discussions, and full texts of original papers.
Models trained on SPICED improve downstream performance on evidence retrieval for fact checking of real-world scientific claims.
arXiv Detail & Related papers (2022-10-24T07:44:38Z) - Generating Scientific Claims for Zero-Shot Scientific Fact Checking [54.62086027306609]
Automated scientific fact checking is difficult due to the complexity of scientific language and a lack of significant amounts of training data.
We propose scientific claim generation, the task of generating one or more atomic and verifiable claims from scientific sentences.
We also demonstrate its usefulness in zero-shot fact checking for biomedical claims.
arXiv Detail & Related papers (2022-03-24T11:29:20Z) - RerrFact: Reduced Evidence Retrieval Representations for Scientific
Claim Verification [4.052777228128475]
We propose a modular approach that sequentially carries out binary classification for every prediction subtask.
We carry out two-step stance predictions that first differentiate non-relevant rationales and then identify supporting or refuting rationales for a given claim.
Experimentally, our system RerrFact with no fine-tuning, simple design, and a fraction of model parameters fairs competitively on the leaderboard.
arXiv Detail & Related papers (2022-02-05T21:52:45Z) - LongChecker: Improving scientific claim verification by modeling
full-abstract context [38.73116177387815]
We introduce the LongChecker system for scientific claim verification.
Given a scientific claim and an evidence-containing research abstract, LongChecker predicts a veracity label and identifies supporting rationales.
By making labeling decisions based on all available context, LongChecker achieves better performance on cases requiring this type of understanding.
arXiv Detail & Related papers (2021-12-02T23:37:16Z) - SciClops: Detecting and Contextualizing Scientific Claims for Assisting
Manual Fact-Checking [7.507186058512835]
This paper describes SciClops, a method to help combat online scientific misinformation.
SciClops involves three main steps to process scientific claims found in online news articles and social media postings.
It effectively assists non-expert fact-checkers in the verification of complex scientific claims, outperforming commercial fact-checking systems.
arXiv Detail & Related papers (2021-10-25T16:35:58Z) - Fact or Fiction: Verifying Scientific Claims [53.29101835904273]
We introduce scientific claim verification, a new task to select abstracts from the research literature containing evidence that SUPPORTS or REFUTES a given scientific claim.
We construct SciFact, a dataset of 1.4K expert-written scientific claims paired with evidence-containing abstracts annotated with labels and rationales.
We show that our system is able to verify claims related to COVID-19 by identifying evidence from the CORD-19 corpus.
arXiv Detail & Related papers (2020-04-30T17:22:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.