NapSS: Paragraph-level Medical Text Simplification via Narrative
Prompting and Sentence-matching Summarization
- URL: http://arxiv.org/abs/2302.05574v1
- Date: Sat, 11 Feb 2023 02:20:25 GMT
- Title: NapSS: Paragraph-level Medical Text Simplification via Narrative
Prompting and Sentence-matching Summarization
- Authors: Junru Lu, Jiazheng Li, Byron C. Wallace, Yulan He, Gabriele Pergola
- Abstract summary: We propose a summarize-then-simplify two-stage strategy, which we call NapSS.
NapSS identifies the relevant content to simplify while ensuring that the original narrative flow is preserved.
Our model achieves significantly better than the seq2seq baseline on an English medical corpus.
- Score: 46.772517928718216
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Accessing medical literature is difficult for laypeople as the content is
written for specialists and contains medical jargon. Automated text
simplification methods offer a potential means to address this issue. In this
work, we propose a summarize-then-simplify two-stage strategy, which we call
NapSS, identifying the relevant content to simplify while ensuring that the
original narrative flow is preserved. In this approach, we first generate
reference summaries via sentence matching between the original and the
simplified abstracts. These summaries are then used to train an extractive
summarizer, learning the most relevant content to be simplified. Then, to
ensure the narrative consistency of the simplified text, we synthesize
auxiliary narrative prompts combining key phrases derived from the syntactical
analyses of the original text. Our model achieves results significantly better
than the seq2seq baseline on an English medical corpus, yielding 3%~4% absolute
improvements in terms of lexical similarity, and providing a further 1.1%
improvement of SARI score when combined with the baseline. We also highlight
shortcomings of existing evaluation methods, and introduce new metrics that
take into account both lexical and high-level semantic similarity. A human
evaluation conducted on a random sample of the test set further establishes the
effectiveness of the proposed approach. Codes and models are released here:
https://github.com/LuJunru/NapSS.
Related papers
- Attribute Structuring Improves LLM-Based Evaluation of Clinical Text
Summaries [62.32403630651586]
Large language models (LLMs) have shown the potential to generate accurate clinical text summaries, but still struggle with issues regarding grounding and evaluation.
Here, we explore a general mitigation framework using Attribute Structuring (AS), which structures the summary evaluation process.
AS consistently improves the correspondence between human annotations and automated metrics in clinical text summarization.
arXiv Detail & Related papers (2024-03-01T21:59:03Z) - AugSumm: towards generalizable speech summarization using synthetic
labels from large language model [61.73741195292997]
Abstractive speech summarization (SSUM) aims to generate human-like summaries from speech.
conventional SSUM models are mostly trained and evaluated with a single ground-truth (GT) human-annotated deterministic summary.
We propose AugSumm, a method to leverage large language models (LLMs) as a proxy for human annotators to generate augmented summaries.
arXiv Detail & Related papers (2024-01-10T18:39:46Z) - Correction with Backtracking Reduces Hallucination in Summarization [29.093092115901694]
Abstractive summarization aims at generating natural language summaries of a source document that are succinct while preserving the important elements.
Despite recent advances, neural text summarization models are known to be susceptible to hallucinating.
We introduce a simple yet efficient technique, CoBa, to reduce hallucination in abstractive summarization.
arXiv Detail & Related papers (2023-10-24T20:48:11Z) - Factually Consistent Summarization via Reinforcement Learning with
Textual Entailment Feedback [57.816210168909286]
We leverage recent progress on textual entailment models to address this problem for abstractive summarization systems.
We use reinforcement learning with reference-free, textual entailment rewards to optimize for factual consistency.
Our results, according to both automatic metrics and human evaluation, show that our method considerably improves the faithfulness, salience, and conciseness of the generated summaries.
arXiv Detail & Related papers (2023-05-31T21:04:04Z) - Comparing Methods for Extractive Summarization of Call Centre Dialogue [77.34726150561087]
We experimentally compare several such methods by using them to produce summaries of calls, and evaluating these summaries objectively.
We found that TopicSum and Lead-N outperform the other summarisation methods, whilst BERTSum received comparatively lower scores in both subjective and objective evaluations.
arXiv Detail & Related papers (2022-09-06T13:16:02Z) - A General Contextualized Rewriting Framework for Text Summarization [15.311467109946571]
Exiting rewriting systems take each extractive sentence as the only input, which is relatively focused but can lose necessary background knowledge and discourse context.
We formalize contextualized rewriting as a seq2seq with group-tag alignments, identifying extractive sentences through content-based addressing.
Results show that our approach significantly outperforms non-contextualized rewriting systems without requiring reinforcement learning.
arXiv Detail & Related papers (2022-07-13T03:55:57Z) - Reinforcing Semantic-Symmetry for Document Summarization [15.113768658584979]
Document summarization condenses a long document into a short version with salient information and accurate semantic descriptions.
This paper introduces a new textbfreinforcing stextbfemantic-textbfsymmetry learning textbfmodel is proposed for document summarization.
A series of experiments have been conducted on two wildly used benchmark datasets CNN/Daily Mail and BigPatent.
arXiv Detail & Related papers (2021-12-14T17:41:37Z) - Text Simplification for Comprehension-based Question-Answering [7.144235435987265]
We release Simple-SQuAD, a simplified version of the widely-used SQuAD dataset.
We benchmark the newly created corpus and perform an ablation study for examining the effect of the simplification process in the SQuAD-based question answering task.
arXiv Detail & Related papers (2021-09-28T18:48:00Z) - Extractive Summarization as Text Matching [123.09816729675838]
This paper creates a paradigm shift with regard to the way we build neural extractive summarization systems.
We formulate the extractive summarization task as a semantic text matching problem.
We have driven the state-of-the-art extractive result on CNN/DailyMail to a new level (44.41 in ROUGE-1)
arXiv Detail & Related papers (2020-04-19T08:27:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.