Salience Estimation with Multi-Attention Learning for Abstractive Text
Summarization
- URL: http://arxiv.org/abs/2004.03589v1
- Date: Tue, 7 Apr 2020 02:38:56 GMT
- Title: Salience Estimation with Multi-Attention Learning for Abstractive Text
Summarization
- Authors: Piji Li, Lidong Bing, Zhongyu Wei, Wai Lam
- Abstract summary: In the task of text summarization, salience estimation for words, phrases or sentences is a critical component.
We propose a Multi-Attention Learning framework which contains two new attention learning components for salience estimation.
- Score: 86.45110800123216
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Attention mechanism plays a dominant role in the sequence generation models
and has been used to improve the performance of machine translation and
abstractive text summarization. Different from neural machine translation, in
the task of text summarization, salience estimation for words, phrases or
sentences is a critical component, since the output summary is a distillation
of the input text. Although the typical attention mechanism can conduct text
fragment selection from the input text conditioned on the decoder states, there
is still a gap to conduct direct and effective salience detection. To bring
back direct salience estimation for summarization with neural networks, we
propose a Multi-Attention Learning framework which contains two new attention
learning components for salience estimation: supervised attention learning and
unsupervised attention learning. We regard the attention weights as the
salience information, which means that the semantic units with large attention
value will be more important. The context information obtained based on the
estimated salience is incorporated with the typical attention mechanism in the
decoder to conduct summary generation. Extensive experiments on some benchmark
datasets in different languages demonstrate the effectiveness of the proposed
framework for the task of abstractive summarization.
Related papers
- Neural Sequence-to-Sequence Modeling with Attention by Leveraging Deep Learning Architectures for Enhanced Contextual Understanding in Abstractive Text Summarization [0.0]
This paper presents a novel framework for abstractive TS of single documents.
It integrates three dominant aspects: structure, semantic, and neural-based approaches.
Results indicate significant improvements in handling rare and OOV words.
arXiv Detail & Related papers (2024-04-08T18:33:59Z) - Attention-aware semantic relevance predicting Chinese sentence reading [6.294658916880712]
This study proposes an attention-aware'' approach for computing contextual semantic relevance.
The attention-aware metrics of semantic relevance can more accurately predict fixation durations in Chinese reading tasks.
Our approach underscores the potential of these metrics to advance our comprehension of how humans understand and process language.
arXiv Detail & Related papers (2024-03-27T13:22:38Z) - Towards Zero-shot Human-Object Interaction Detection via Vision-Language
Integration [14.678931157058363]
We propose a novel framework, termed Knowledge Integration to HOI (KI2HOI), that effectively integrates the knowledge of visual-language model to improve zero-shot HOI detection.
We develop an effective additive self-attention mechanism to generate more comprehensive visual representations.
Our model outperforms the previous methods in various zero-shot and full-supervised settings.
arXiv Detail & Related papers (2024-03-12T02:07:23Z) - Picking the Underused Heads: A Network Pruning Perspective of Attention
Head Selection for Fusing Dialogue Coreference Information [50.41829484199252]
Transformer-based models with the multi-head self-attention mechanism are widely used in natural language processing.
We investigate the attention head selection and manipulation strategy for feature injection from a network pruning perspective.
arXiv Detail & Related papers (2023-12-15T05:27:24Z) - Robust Saliency-Aware Distillation for Few-shot Fine-grained Visual
Recognition [57.08108545219043]
Recognizing novel sub-categories with scarce samples is an essential and challenging research topic in computer vision.
Existing literature addresses this challenge by employing local-based representation approaches.
This article proposes a novel model, Robust Saliency-aware Distillation (RSaD), for few-shot fine-grained visual recognition.
arXiv Detail & Related papers (2023-05-12T00:13:17Z) - LadRa-Net: Locally-Aware Dynamic Re-read Attention Net for Sentence
Semantic Matching [66.65398852962177]
We develop a novel Dynamic Re-read Network (DRr-Net) for sentence semantic matching.
We extend DRr-Net to Locally-Aware Dynamic Re-read Attention Net (LadRa-Net)
Experiments on two popular sentence semantic matching tasks demonstrate that DRr-Net can significantly improve the performance of sentence semantic matching.
arXiv Detail & Related papers (2021-08-06T02:07:04Z) - Narrative Incoherence Detection [76.43894977558811]
We propose the task of narrative incoherence detection as a new arena for inter-sentential semantic understanding.
Given a multi-sentence narrative, decide whether there exist any semantic discrepancies in the narrative flow.
arXiv Detail & Related papers (2020-12-21T07:18:08Z) - Few-Shot Learning for Opinion Summarization [117.70510762845338]
Opinion summarization is the automatic creation of text reflecting subjective information expressed in multiple documents.
In this work, we show that even a handful of summaries is sufficient to bootstrap generation of the summary text.
Our approach substantially outperforms previous extractive and abstractive methods in automatic and human evaluation.
arXiv Detail & Related papers (2020-04-30T15:37:38Z) - Interpretable Multi-Headed Attention for Abstractive Summarization at
Controllable Lengths [14.762731718325002]
Multi-level Summarizer (MLS) is a supervised method to construct abstractive summaries of a text document at controllable lengths.
MLS outperforms strong baselines by up to 14.70% in the METEOR score.
arXiv Detail & Related papers (2020-02-18T19:40:20Z) - Read Beyond the Lines: Understanding the Implied Textual Meaning via a
Skim and Intensive Reading Model [41.61803103143516]
We propose a novel, simple, and effective deep neural framework, called Skim and Intensive Reading Model (SIRM)
The proposed SIRM consists of two main components, namely the skim reading component and intensive reading component.
We conduct extensive comparative experiments on several sarcasm benchmarks and an industrial spam dataset with metaphors.
arXiv Detail & Related papers (2020-01-03T03:43:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.