Biased TextRank: Unsupervised Graph-Based Content Extraction
- URL: http://arxiv.org/abs/2011.01026v1
- Date: Mon, 2 Nov 2020 15:17:44 GMT
- Title: Biased TextRank: Unsupervised Graph-Based Content Extraction
- Authors: Ashkan Kazemi, Ver\'onica P\'erez-Rosas, Rada Mihalcea
- Abstract summary: Biased TextRank is a graph-based content extraction method inspired by the popular TextRank algorithm.
We present two applications of Biased TextRank: focused summarization and explanation extraction.
- Score: 26.54218341713572
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce Biased TextRank, a graph-based content extraction method
inspired by the popular TextRank algorithm that ranks text spans according to
their importance for language processing tasks and according to their relevance
to an input "focus." Biased TextRank enables focused content extraction for
text by modifying the random restarts in the execution of TextRank. The random
restart probabilities are assigned based on the relevance of the graph nodes to
the focus of the task. We present two applications of Biased TextRank: focused
summarization and explanation extraction, and show that our algorithm leads to
improved performance on two different datasets by significant ROUGE-N score
margins. Much like its predecessor, Biased TextRank is unsupervised, easy to
implement and orders of magnitude faster and lighter than current
state-of-the-art Natural Language Processing methods for similar tasks.
Related papers
- DNTextSpotter: Arbitrary-Shaped Scene Text Spotting via Improved Denoising Training [17.734265617973293]
We propose a novel denoising training method (DNTextSpotter) for arbitrary-shaped text spotting.
DNTextSpotter decomposes the queries of the denoising part into noised positional queries and noised content queries.
It outperforms the state-of-the-art methods on four benchmarks.
arXiv Detail & Related papers (2024-08-01T07:52:07Z) - Copy Is All You Need [66.00852205068327]
We formulate text generation as progressively copying text segments from an existing text collection.
Our approach achieves better generation quality according to both automatic and human evaluations.
Our approach attains additional performance gains by simply scaling up to larger text collections.
arXiv Detail & Related papers (2023-07-13T05:03:26Z) - TextFormer: A Query-based End-to-End Text Spotter with Mixed Supervision [61.186488081379]
We propose TextFormer, a query-based end-to-end text spotter with Transformer architecture.
TextFormer builds upon an image encoder and a text decoder to learn a joint semantic understanding for multi-task modeling.
It allows for mutual training and optimization of classification, segmentation, and recognition branches, resulting in deeper feature sharing.
arXiv Detail & Related papers (2023-06-06T03:37:41Z) - Evaluating Factual Consistency of Texts with Semantic Role Labeling [3.1776833268555134]
We introduce SRLScore, a reference-free evaluation metric designed with text summarization in mind.
A final factuality score is computed by an adjustable scoring mechanism.
Correlation with human judgments on English summarization datasets shows that SRLScore is competitive with state-of-the-art methods.
arXiv Detail & Related papers (2023-05-22T17:59:42Z) - ChatGraph: Interpretable Text Classification by Converting ChatGPT
Knowledge to Graphs [54.48467003509595]
ChatGPT has shown superior performance in various natural language processing (NLP) tasks.
We propose a novel framework that leverages the power of ChatGPT for specific tasks, such as text classification.
Our method provides a more transparent decision-making process compared with previous text classification methods.
arXiv Detail & Related papers (2023-05-03T19:57:43Z) - Graph-based Semantical Extractive Text Analysis [0.0]
In this work, we improve the results of the TextRank algorithm by incorporating the semantic similarity between parts of the text.
Aside from keyword extraction and text summarization, we develop a topic clustering algorithm based on our framework.
arXiv Detail & Related papers (2022-12-19T18:30:26Z) - Text Revision by On-the-Fly Representation Optimization [76.11035270753757]
Current state-of-the-art methods formulate these tasks as sequence-to-sequence learning problems.
We present an iterative in-place editing approach for text revision, which requires no parallel data.
It achieves competitive and even better performance than state-of-the-art supervised methods on text simplification.
arXiv Detail & Related papers (2022-04-15T07:38:08Z) - Be More with Less: Hypergraph Attention Networks for Inductive Text
Classification [56.98218530073927]
Graph neural networks (GNNs) have received increasing attention in the research community and demonstrated their promising results on this canonical task.
Despite the success, their performance could be largely jeopardized in practice since they are unable to capture high-order interaction between words.
We propose a principled model -- hypergraph attention networks (HyperGAT) which can obtain more expressive power with less computational consumption for text representation learning.
arXiv Detail & Related papers (2020-11-01T00:21:59Z) - Rank over Class: The Untapped Potential of Ranking in Natural Language
Processing [8.637110868126546]
We argue that many tasks which are currently addressed using classification are in fact being shoehorned into a classification mould.
We propose a novel end-to-end ranking approach consisting of a Transformer network responsible for producing representations for a pair of text sequences.
In an experiment on a heavily-skewed sentiment analysis dataset, converting ranking results to classification labels yields an approximately 22% improvement over state-of-the-art text classification.
arXiv Detail & Related papers (2020-09-10T22:18:57Z) - Abstractive Summarization of Spoken and Written Instructions with BERT [66.14755043607776]
We present the first application of the BERTSum model to conversational language.
We generate abstractive summaries of narrated instructional videos across a wide variety of topics.
We envision this integrated as a feature in intelligent virtual assistants, enabling them to summarize both written and spoken instructional content upon request.
arXiv Detail & Related papers (2020-08-21T20:59:34Z) - Investigating Label Bias in Beam Search for Open-ended Text Generation [8.331919991368366]
In open-ended text generation, beam search is often found to produce repetitive and generic texts.
Standard seq2seq models suffer from label bias due to its locally normalized probability formulation.
By combining locally normalized maximum likelihood estimation and globally normalized sequence-level training, label bias can be reduced with almost no sacrifice in perplexity.
arXiv Detail & Related papers (2020-05-22T05:17:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.