Abstractive Text Summarization using Attentive GRU based Encoder-Decoder
- URL: http://arxiv.org/abs/2302.13117v1
- Date: Sat, 25 Feb 2023 16:45:46 GMT
- Title: Abstractive Text Summarization using Attentive GRU based Encoder-Decoder
- Authors: Tohida Rehman, Suchandan Das, Debarshi Kumar Sanyal, Samiran
Chattopadhyay
- Abstract summary: Automatic text summarization has emerged as an important application of machine learning in text processing.
In this paper, an english text summarizer has been built with GRU-based encoder and decoder.
The output is observed to outperform competitive models in the literature.
- Score: 4.339043862780233
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In todays era huge volume of information exists everywhere. Therefore, it is
very crucial to evaluate that information and extract useful, and often
summarized, information out of it so that it may be used for relevant purposes.
This extraction can be achieved through a crucial technique of artificial
intelligence, namely, machine learning. Indeed automatic text summarization has
emerged as an important application of machine learning in text processing. In
this paper, an english text summarizer has been built with GRU-based encoder
and decoder. Bahdanau attention mechanism has been added to overcome the
problem of handling long sequences in the input text. A news-summary dataset
has been used to train the model. The output is observed to outperform
competitive models in the literature. The generated summary can be used as a
newspaper headline.
Related papers
- Text2Data: Low-Resource Data Generation with Textual Control [100.5970757736845]
Text2Data is a novel approach that utilizes unlabeled data to understand the underlying data distribution.
It undergoes finetuning via a novel constraint optimization-based learning objective that ensures controllability and effectively counteracts catastrophic forgetting.
arXiv Detail & Related papers (2024-02-08T03:41:39Z) - Optimizing Factual Accuracy in Text Generation through Dynamic Knowledge
Selection [71.20871905457174]
Language models (LMs) have revolutionized the way we interact with information, but they often generate nonfactual text.
Previous methods use external knowledge as references for text generation to enhance factuality but often struggle with the knowledge mix-up of irrelevant references.
We present DKGen, which divide the text generation process into an iterative process.
arXiv Detail & Related papers (2023-08-30T02:22:40Z) - Abstractive Summary Generation for the Urdu Language [1.9594639581421422]
We employ a transformer-based model that utilizes self-attention mechanisms to encode the input text and generate a summary.
Our experiments show that our model can produce summaries that are grammatically correct and semantically meaningful.
arXiv Detail & Related papers (2023-05-25T15:55:42Z) - Tram: A Token-level Retrieval-augmented Mechanism for Source Code Summarization [76.57699934689468]
We propose a fine-grained Token-level retrieval-augmented mechanism (Tram) on the decoder side to enhance the performance of neural models.
To overcome the challenge of token-level retrieval in capturing contextual code semantics, we also propose integrating code semantics into individual summary tokens.
arXiv Detail & Related papers (2023-05-18T16:02:04Z) - Automated News Summarization Using Transformers [4.932130498861987]
We will be presenting a comprehensive comparison of a few transformer architecture based pre-trained models for text summarization.
For analysis and comparison, we have used the BBC news dataset that contains text data that can be used for summarization and human generated summaries.
arXiv Detail & Related papers (2021-04-23T04:22:33Z) - TEACHTEXT: CrossModal Generalized Distillation for Text-Video Retrieval [103.85002875155551]
We propose a novel generalized distillation method, TeachText, for exploiting large-scale language pretraining.
We extend our method to video side modalities and show that we can effectively reduce the number of used modalities at test time.
Our approach advances the state of the art on several video retrieval benchmarks by a significant margin and adds no computational overhead at test time.
arXiv Detail & Related papers (2021-04-16T17:55:28Z) - A Survey on Semi-parametric Machine Learning Technique for Time Series
Forecasting [4.9341230675162215]
Grey Machine Learning (GML) is capable of handling large datasets as well as small datasets for time series forecasting likely outcomes.
This survey presents a comprehensive overview of the existing semi-parametric machine learning techniques for time series forecasting.
arXiv Detail & Related papers (2021-04-02T03:26:20Z) - A Transformer-based Approach for Source Code Summarization [86.08359401867577]
We learn code representation for summarization by modeling the pairwise relationship between code tokens.
We show that despite the approach is simple, it outperforms the state-of-the-art techniques by a significant margin.
arXiv Detail & Related papers (2020-05-01T23:29:36Z) - Salience Estimation with Multi-Attention Learning for Abstractive Text
Summarization [86.45110800123216]
In the task of text summarization, salience estimation for words, phrases or sentences is a critical component.
We propose a Multi-Attention Learning framework which contains two new attention learning components for salience estimation.
arXiv Detail & Related papers (2020-04-07T02:38:56Z) - Hybrid Attention-Based Transformer Block Model for Distant Supervision
Relation Extraction [20.644215991166902]
We propose a new framework using hybrid attention-based Transformer block with multi-instance learning to perform the DSRE task.
The proposed approach can outperform the state-of-the-art algorithms on the evaluation dataset.
arXiv Detail & Related papers (2020-03-10T13:05:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.