Balancing Lexical and Semantic Quality in Abstractive Summarization
- URL: http://arxiv.org/abs/2305.09898v1
- Date: Wed, 17 May 2023 02:18:31 GMT
- Title: Balancing Lexical and Semantic Quality in Abstractive Summarization
- Authors: Jeewoo Sul and Yong Suk Choi
- Abstract summary: We propose a novel training method in which a re-ranker balances the lexical and semantic quality.
Experiments on the CNN/DailyMail and XSum datasets show that our method can estimate the meaning of summaries without seriously degrading the lexical aspect.
- Score: 0.38073142980733
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: An important problem of the sequence-to-sequence neural models widely used in
abstractive summarization is exposure bias. To alleviate this problem,
re-ranking systems have been applied in recent years. Despite some performance
improvements, this approach remains underexplored. Previous works have mostly
specified the rank through the ROUGE score and aligned candidate summaries, but
there can be quite a large gap between the lexical overlap metric and semantic
similarity. In this paper, we propose a novel training method in which a
re-ranker balances the lexical and semantic quality. We further newly define
false positives in ranking and present a strategy to reduce their influence.
Experiments on the CNN/DailyMail and XSum datasets show that our method can
estimate the meaning of summaries without seriously degrading the lexical
aspect. More specifically, it achieves an 89.67 BERTScore on the CNN/DailyMail
dataset, reaching new state-of-the-art performance. Our code is publicly
available at https://github.com/jeewoo1025/BalSum.
Related papers
- Investigating Text Shortening Strategy in BERT: Truncation vs Summarization [2.7645945793246973]
This study investigates the performance of document truncation and summarization in text classification tasks.
We used a dataset of summarization tasks based on Indonesian news articles to do classification tests.
arXiv Detail & Related papers (2024-03-19T15:01:14Z) - Enhancing Consistency and Mitigating Bias: A Data Replay Approach for
Incremental Learning [100.7407460674153]
Deep learning systems are prone to catastrophic forgetting when learning from a sequence of tasks.
To mitigate the problem, a line of methods propose to replay the data of experienced tasks when learning new tasks.
However, it is not expected in practice considering the memory constraint or data privacy issue.
As a replacement, data-free data replay methods are proposed by inverting samples from the classification model.
arXiv Detail & Related papers (2024-01-12T12:51:12Z) - On Context Utilization in Summarization with Large Language Models [83.84459732796302]
Large language models (LLMs) excel in abstractive summarization tasks, delivering fluent and pertinent summaries.
Recent advancements have extended their capabilities to handle long-input contexts, exceeding 100k tokens.
We conduct the first comprehensive study on context utilization and position bias in summarization.
arXiv Detail & Related papers (2023-10-16T16:45:12Z) - Evaluating Graph Neural Networks for Link Prediction: Current Pitfalls
and New Benchmarking [66.83273589348758]
Link prediction attempts to predict whether an unseen edge exists based on only a portion of edges of a graph.
A flurry of methods have been introduced in recent years that attempt to make use of graph neural networks (GNNs) for this task.
New and diverse datasets have also been created to better evaluate the effectiveness of these new models.
arXiv Detail & Related papers (2023-06-18T01:58:59Z) - Towards Summary Candidates Fusion [26.114829566197976]
We propose a new paradigm in second-stage abstractive summarization called SummaFusion.
It fuses several summary candidates to produce a novel abstractive second-stage summary.
Our method works well on several summarization datasets, improving both the ROUGE scores and qualitative properties of fused summaries.
arXiv Detail & Related papers (2022-10-17T06:48:05Z) - Entity Disambiguation with Entity Definitions [50.01142092276296]
Local models have recently attained astounding performances in Entity Disambiguation (ED)
Previous works limited their studies to using, as the textual representation of each candidate, only its Wikipedia title.
In this paper, we address this limitation and investigate to what extent more expressive textual representations can mitigate it.
We report a new state of the art on 2 out of 6 benchmarks we consider and strongly improve the generalization capability over unseen patterns.
arXiv Detail & Related papers (2022-10-11T17:46:28Z) - Semantic-Preserving Adversarial Text Attacks [85.32186121859321]
We propose a Bigram and Unigram based adaptive Semantic Preservation Optimization (BU-SPO) method to examine the vulnerability of deep models.
Our method achieves the highest attack success rates and semantics rates by changing the smallest number of words compared with existing methods.
arXiv Detail & Related papers (2021-08-23T09:05:18Z) - Unsupervised Extractive Summarization by Pre-training Hierarchical
Transformers [107.12125265675483]
Unsupervised extractive document summarization aims to select important sentences from a document without using labeled summaries during training.
Existing methods are mostly graph-based with sentences as nodes and edge weights measured by sentence similarities.
We find that transformer attentions can be used to rank sentences for unsupervised extractive summarization.
arXiv Detail & Related papers (2020-10-16T08:44:09Z) - Rank over Class: The Untapped Potential of Ranking in Natural Language
Processing [8.637110868126546]
We argue that many tasks which are currently addressed using classification are in fact being shoehorned into a classification mould.
We propose a novel end-to-end ranking approach consisting of a Transformer network responsible for producing representations for a pair of text sequences.
In an experiment on a heavily-skewed sentiment analysis dataset, converting ranking results to classification labels yields an approximately 22% improvement over state-of-the-art text classification.
arXiv Detail & Related papers (2020-09-10T22:18:57Z) - PushNet: Efficient and Adaptive Neural Message Passing [1.9121961872220468]
Message passing neural networks have recently evolved into a state-of-the-art approach to representation learning on graphs.
Existing methods perform synchronous message passing along all edges in multiple subsequent rounds.
We consider a novel asynchronous message passing approach where information is pushed only along the most relevant edges until convergence.
arXiv Detail & Related papers (2020-03-04T18:15:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.