LED down the rabbit hole: exploring the potential of global attention
for biomedical multi-document summarisation
- URL: http://arxiv.org/abs/2209.08698v1
- Date: Mon, 19 Sep 2022 01:13:42 GMT
- Title: LED down the rabbit hole: exploring the potential of global attention
for biomedical multi-document summarisation
- Authors: Yulia Otmakhova, Hung Thinh Truong, Timothy Baldwin, Trevor Cohn,
Karin Verspoor, Jey Han Lau
- Abstract summary: We adapt PRIMERA to the biomedical domain by placing global attention on important biomedical entities.
We analyse the outputs of the 23 resulting models, and report patterns in the results related to the presence of additional global attention.
- Score: 59.307534363825816
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: In this paper we report on our submission to the Multidocument Summarisation
for Literature Review (MSLR) shared task. Specifically, we adapt PRIMERA (Xiao
et al., 2022) to the biomedical domain by placing global attention on important
biomedical entities in several ways. We analyse the outputs of the 23 resulting
models, and report patterns in the results related to the presence of
additional global attention, number of training steps, and the input
configuration.
Related papers
- A Refer-and-Ground Multimodal Large Language Model for Biomedicine [10.519866875035003]
The Med-GRIT-270k dataset is the first dedicated to the biomedical domain and integrates refer and ground conversations.
We introduce a Refer-and-Ground Multimodal Large Language Model for Biomedicine (BiRD) by using this dataset and multi-task instruction learning.
arXiv Detail & Related papers (2024-06-26T07:56:17Z) - Intent Detection and Entity Extraction from BioMedical Literature [14.52164637112797]
Large-language models (LLMs) motivated by endeavours to attain generalized intelligence, their efficacy in replacing task and domain-specific natural language understanding approaches remains questionable.
We show that Supervised Fine Tuned approaches are still relevant and more effective than general-purpose LLMs.
arXiv Detail & Related papers (2024-04-04T17:09:52Z) - An Evaluation of Large Language Models in Bioinformatics Research [52.100233156012756]
We study the performance of large language models (LLMs) on a wide spectrum of crucial bioinformatics tasks.
These tasks include the identification of potential coding regions, extraction of named entities for genes and proteins, detection of antimicrobial and anti-cancer peptides, molecular optimization, and resolution of educational bioinformatics problems.
Our findings indicate that, given appropriate prompts, LLMs like GPT variants can successfully handle most of these tasks.
arXiv Detail & Related papers (2024-02-21T11:27:31Z) - BioLORD-2023: Semantic Textual Representations Fusing LLM and Clinical
Knowledge Graph Insights [15.952942443163474]
We propose a new state-of-the-art approach for obtaining high-fidelity representations of biomedical concepts and sentences.
We demonstrate consistent and substantial performance improvements over the previous state of the art.
Besides our new state-of-the-art biomedical model for English, we also distill and release a multilingual model compatible with 50+ languages.
arXiv Detail & Related papers (2023-11-27T18:46:17Z) - Overview of the BioLaySumm 2023 Shared Task on Lay Summarization of
Biomedical Research Articles [47.04555835353173]
This paper presents the results of the shared task on Lay Summarisation of Biomedical Research Articles (BioLaySumm) hosted at the BioNLP Workshop at ACL 2023.
The goal of this shared task is to develop abstractive summarisation models capable of generating "lay summaries"
In addition to overall results, we report on the setup and insights from the BioLaySumm shared task, which attracted a total of 20 participating teams across both subtasks.
arXiv Detail & Related papers (2023-09-29T15:43:42Z) - PMC-LLaMA: Towards Building Open-source Language Models for Medicine [62.39105735933138]
Large Language Models (LLMs) have showcased remarkable capabilities in natural language understanding.
LLMs struggle in domains that require precision, such as medical applications, due to their lack of domain-specific knowledge.
We describe the procedure for building a powerful, open-source language model specifically designed for medicine applications, termed as PMC-LLaMA.
arXiv Detail & Related papers (2023-04-27T18:29:05Z) - Discovering Drug-Target Interaction Knowledge from Biomedical Literature [107.98712673387031]
The Interaction between Drugs and Targets (DTI) in human body plays a crucial role in biomedical science and applications.
As millions of papers come out every year in the biomedical domain, automatically discovering DTI knowledge from literature becomes an urgent demand in the industry.
We explore the first end-to-end solution for this task by using generative approaches.
We regard the DTI triplets as a sequence and use a Transformer-based model to directly generate them without using the detailed annotations of entities and relations.
arXiv Detail & Related papers (2021-09-27T17:00:14Z) - Domain-Specific Pretraining for Vertical Search: Case Study on
Biomedical Literature [67.4680600632232]
Self-supervised learning has emerged as a promising direction to overcome the annotation bottleneck.
We propose a general approach for vertical search based on domain-specific pretraining.
Our system can scale to tens of millions of articles on PubMed and has been deployed as Microsoft Biomedical Search.
arXiv Detail & Related papers (2021-06-25T01:02:55Z) - MS2: Multi-Document Summarization of Medical Studies [11.38740406132287]
We release MS2 (Multi-Document Summarization of Medical Studies), a dataset of over 470k documents and 20k summaries derived from the scientific literature.
This dataset facilitates the development of systems that can assess and aggregate contradictory evidence across multiple studies.
We experiment with a summarization system based on BART, with promising early results.
arXiv Detail & Related papers (2021-04-13T19:59:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.