Neural Rankers for Effective Screening Prioritisation in Medical
Systematic Review Literature Search
- URL: http://arxiv.org/abs/2212.09017v1
- Date: Sun, 18 Dec 2022 05:26:40 GMT
- Title: Neural Rankers for Effective Screening Prioritisation in Medical
Systematic Review Literature Search
- Authors: Shuai Wang, Harrisen Scells, Bevan Koopman, Guido Zuccon
- Abstract summary: We apply several pre-trained language models to the systematic review document ranking task.
An empirical analysis compares how effective neural methods compare to traditional methods for this task.
Our results show that BERT-based rankers outperform the current state-of-the-art screening prioritisation methods.
- Score: 31.797257552928336
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Medical systematic reviews typically require assessing all the documents
retrieved by a search. The reason is two-fold: the task aims for ``total
recall''; and documents retrieved using Boolean search are an unordered set,
and thus it is unclear how an assessor could examine only a subset. Screening
prioritisation is the process of ranking the (unordered) set of retrieved
documents, allowing assessors to begin the downstream processes of the
systematic review creation earlier, leading to earlier completion of the
review, or even avoiding screening documents ranked least relevant.
Screening prioritisation requires highly effective ranking methods.
Pre-trained language models are state-of-the-art on many IR tasks but have yet
to be applied to systematic review screening prioritisation. In this paper, we
apply several pre-trained language models to the systematic review document
ranking task, both directly and fine-tuned. An empirical analysis compares how
effective neural methods compare to traditional methods for this task. We also
investigate different types of document representations for neural methods and
their impact on ranking performance.
Our results show that BERT-based rankers outperform the current
state-of-the-art screening prioritisation methods. However, BERT rankers and
existing methods can actually be complementary, and thus, further improvements
may be achieved if used in conjunction.
Related papers
- JudgeRank: Leveraging Large Language Models for Reasoning-Intensive Reranking [81.88787401178378]
We introduce JudgeRank, a novel agentic reranker that emulates human cognitive processes when assessing document relevance.
We evaluate JudgeRank on the reasoning-intensive BRIGHT benchmark, demonstrating substantial performance improvements over first-stage retrieval methods.
In addition, JudgeRank performs on par with fine-tuned state-of-the-art rerankers on the popular BEIR benchmark, validating its zero-shot generalization capability.
arXiv Detail & Related papers (2024-10-31T18:43:12Z) - Generating Natural Language Queries for More Effective Systematic Review
Screening Prioritisation [53.77226503675752]
The current state of the art uses the final title of the review as a query to rank the documents using BERT-based neural rankers.
In this paper, we explore alternative sources of queries for prioritising screening, such as the Boolean query used to retrieve the documents to be screened and queries generated by instruction-based large-scale language models such as ChatGPT and Alpaca.
Our best approach is not only viable based on the information available at the time of screening, but also has similar effectiveness to the final title.
arXiv Detail & Related papers (2023-09-11T05:12:14Z) - GERE: Generative Evidence Retrieval for Fact Verification [57.78768817972026]
We propose GERE, the first system that retrieves evidences in a generative fashion.
The experimental results on the FEVER dataset show that GERE achieves significant improvements over the state-of-the-art baselines.
arXiv Detail & Related papers (2022-04-12T03:49:35Z) - Towards Reducing Manual Workload in Technology-Assisted Reviews:
Estimating Ranking Performance [30.29371206568408]
When researchers label studies, they can screen ranked documents where relevant documents are higher than irrelevant ones.
This paper investigates the quality of document ranking of systematic reviews.
After extensive analysis on SR document rankings, we hypothesize 'topic broadness' as a factor that affects the ranking quality of SR.
arXiv Detail & Related papers (2022-01-14T19:48:45Z) - Mirror Matching: Document Matching Approach in Seed-driven Document
Ranking for Medical Systematic Reviews [31.3220495275256]
Document ranking is an approach for assisting researchers by providing document rankings where relevant documents are ranked higher than irrelevant ones.
We propose a document matching measure named Mirror Matching, which calculates matching scores between medical abstract texts by incorporating common writing patterns.
arXiv Detail & Related papers (2021-12-28T22:27:52Z) - CODER: An efficient framework for improving retrieval through
COntextualized Document Embedding Reranking [11.635294568328625]
We present a framework for improving the performance of a wide class of retrieval models at minimal computational cost.
It utilizes precomputed document representations extracted by a base dense retrieval method.
It incurs a negligible computational overhead on top of any first-stage method at run time, allowing it to be easily combined with any state-of-the-art dense retrieval method.
arXiv Detail & Related papers (2021-12-16T10:25:26Z) - Value Retrieval with Arbitrary Queries for Form-like Documents [50.5532781148902]
We propose value retrieval with arbitrary queries for form-like documents.
Our method predicts target value for an arbitrary query based on the understanding of layout and semantics of a form.
We propose a simple document language modeling (simpleDLM) strategy to improve document understanding on large-scale model pre-training.
arXiv Detail & Related papers (2021-12-15T01:12:02Z) - Automating Document Classification with Distant Supervision to Increase
the Efficiency of Systematic Reviews [18.33687903724145]
Well-done systematic reviews are expensive, time-demanding, and labor-intensive.
We propose an automatic document classification approach to significantly reduce the effort in reviewing documents.
arXiv Detail & Related papers (2020-12-09T22:45:40Z) - Document Ranking with a Pretrained Sequence-to-Sequence Model [56.44269917346376]
We show how a sequence-to-sequence model can be trained to generate relevance labels as "target words"
Our approach significantly outperforms an encoder-only model in a data-poor regime.
arXiv Detail & Related papers (2020-03-14T22:29:50Z) - Pre-training Tasks for Embedding-based Large-scale Retrieval [68.01167604281578]
We consider the large-scale query-document retrieval problem.
Given a query (e.g., a question), return the set of relevant documents from a large document corpus.
We show that the key ingredient of learning a strong embedding-based Transformer model is the set of pre-training tasks.
arXiv Detail & Related papers (2020-02-10T16:44:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.