Semantic Extractor-Paraphraser based Abstractive Summarization
- URL: http://arxiv.org/abs/2105.01296v1
- Date: Tue, 4 May 2021 05:24:28 GMT
- Title: Semantic Extractor-Paraphraser based Abstractive Summarization
- Authors: Anubhav Jangra, Raghav Jain, Vaibhav Mavi, Sriparna Saha, Pushpak
Bhattacharyya
- Abstract summary: We propose an extractor-paraphraser based abstractive summarization system that exploits semantic overlap.
Our model outperforms the state-of-the-art baselines in terms of ROUGE, METEOR and word similarity (WMS)
- Score: 40.05739160204135
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The anthology of spoken languages today is inundated with textual
information, necessitating the development of automatic summarization models.
In this manuscript, we propose an extractor-paraphraser based abstractive
summarization system that exploits semantic overlap as opposed to its
predecessors that focus more on syntactic information overlap. Our model
outperforms the state-of-the-art baselines in terms of ROUGE, METEOR and word
mover similarity (WMS), establishing the superiority of the proposed system via
extensive ablation experiments. We have also challenged the summarization
capabilities of the state of the art Pointer Generator Network (PGN), and
through thorough experimentation, shown that PGN is more of a paraphraser,
contrary to the prevailing notion of a summarizer; illustrating it's
incapability to accumulate information across multiple sentences.
Related papers
- Leveraging Discourse Structure for Extractive Meeting Summarization [26.76383031532945]
We introduce an extractive summarization system for meetings that leverages discourse structure to better identify salient information.
We train a GNN-based node classification model to select the most important utterances, which are then combined to create an extractive summary.
Experimental results on AMI and ICSI demonstrate that our approach surpasses existing text-based and graph-based extractive summarization systems.
arXiv Detail & Related papers (2024-05-17T19:06:20Z) - Neural Sequence-to-Sequence Modeling with Attention by Leveraging Deep Learning Architectures for Enhanced Contextual Understanding in Abstractive Text Summarization [0.0]
This paper presents a novel framework for abstractive TS of single documents.
It integrates three dominant aspects: structure, semantic, and neural-based approaches.
Results indicate significant improvements in handling rare and OOV words.
arXiv Detail & Related papers (2024-04-08T18:33:59Z) - Robust Saliency-Aware Distillation for Few-shot Fine-grained Visual
Recognition [57.08108545219043]
Recognizing novel sub-categories with scarce samples is an essential and challenging research topic in computer vision.
Existing literature addresses this challenge by employing local-based representation approaches.
This article proposes a novel model, Robust Saliency-aware Distillation (RSaD), for few-shot fine-grained visual recognition.
arXiv Detail & Related papers (2023-05-12T00:13:17Z) - The Current State of Summarization [0.0]
This work aims to concisely indicate the current state of the art in abstractive text summarization.
We outline the current paradigm shifts towards pre-trained encoder-decoder models and large autoregressive language models.
arXiv Detail & Related papers (2023-05-08T17:00:49Z) - Unsupervised Syntactically Controlled Paraphrase Generation with
Abstract Meaning Representations [59.10748929158525]
Abstract Representations (AMR) can greatly improve the performance of unsupervised syntactically controlled paraphrase generation.
Our proposed model, AMR-enhanced Paraphrase Generator (AMRPG), encodes the AMR graph and the constituency parses the input sentence into two disentangled semantic and syntactic embeddings.
Experiments show that AMRPG generates more accurate syntactically controlled paraphrases, both quantitatively and qualitatively, compared to the existing unsupervised approaches.
arXiv Detail & Related papers (2022-11-02T04:58:38Z) - Transition-based Abstract Meaning Representation Parsing with Contextual
Embeddings [0.0]
We study a way of combing two of the most successful routes to meaning of language--statistical language models and symbolic semantics formalisms--in the task of semantic parsing.
We explore the utility of incorporating pretrained context-aware word embeddings--such as BERT and RoBERTa--in the problem of parsing.
arXiv Detail & Related papers (2022-06-13T15:05:24Z) - Attention-based Multi-hypothesis Fusion for Speech Summarization [83.04957603852571]
Speech summarization can be achieved by combining automatic speech recognition (ASR) and text summarization (TS)
ASR errors directly affect the quality of the output summary in the cascade approach.
We propose a cascade speech summarization model that is robust to ASR errors and that exploits multiple hypotheses generated by ASR to attenuate the effect of ASR errors on the summary.
arXiv Detail & Related papers (2021-11-16T03:00:29Z) - Introducing Syntactic Structures into Target Opinion Word Extraction
with Deep Learning [89.64620296557177]
We propose to incorporate the syntactic structures of the sentences into the deep learning models for targeted opinion word extraction.
We also introduce a novel regularization technique to improve the performance of the deep learning models.
The proposed model is extensively analyzed and achieves the state-of-the-art performance on four benchmark datasets.
arXiv Detail & Related papers (2020-10-26T07:13:17Z) - Topic-Guided Abstractive Text Summarization: a Joint Learning Approach [19.623946402970933]
We introduce a new approach for abstractive text summarization, Topic-Guided Abstractive Summarization.
The idea is to incorporate neural topic modeling with a Transformer-based sequence-to-sequence (seq2seq) model in a joint learning framework.
arXiv Detail & Related papers (2020-10-20T14:45:25Z) - Multi-Fact Correction in Abstractive Text Summarization [98.27031108197944]
Span-Fact is a suite of two factual correction models that leverages knowledge learned from question answering models to make corrections in system-generated summaries via span selection.
Our models employ single or multi-masking strategies to either iteratively or auto-regressively replace entities in order to ensure semantic consistency w.r.t. the source text.
Experiments show that our models significantly boost the factual consistency of system-generated summaries without sacrificing summary quality in terms of both automatic metrics and human evaluation.
arXiv Detail & Related papers (2020-10-06T02:51:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.