Summarizing Text on Any Aspects: A Knowledge-Informed Weakly-Supervised
Approach
- URL: http://arxiv.org/abs/2010.06792v2
- Date: Sun, 18 Oct 2020 08:58:23 GMT
- Title: Summarizing Text on Any Aspects: A Knowledge-Informed Weakly-Supervised
Approach
- Authors: Bowen Tan, Lianhui Qin, Eric P. Xing, Zhiting Hu
- Abstract summary: We study summarizing on arbitrary aspects relevant to the document.
Due to the lack of supervision data, we develop a new weak supervision construction method and an aspect modeling scheme.
Experiments show our approach achieves performance boosts on summarizing both real and synthetic documents.
- Score: 89.56158561087209
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Given a document and a target aspect (e.g., a topic of interest),
aspect-based abstractive summarization attempts to generate a summary with
respect to the aspect. Previous studies usually assume a small pre-defined set
of aspects and fall short of summarizing on other diverse topics. In this work,
we study summarizing on arbitrary aspects relevant to the document, which
significantly expands the application of the task in practice. Due to the lack
of supervision data, we develop a new weak supervision construction method and
an aspect modeling scheme, both of which integrate rich external knowledge
sources such as ConceptNet and Wikipedia. Experiments show our approach
achieves performance boosts on summarizing both real and synthetic documents
given pre-defined or arbitrary aspects.
Related papers
- Abstractive Text Summarization: State of the Art, Challenges, and Improvements [6.349503549199403]
This review takes a comprehensive approach encompassing state-of-the-art methods, challenges, solutions, comparisons, limitations and charts out future improvements.
The paper highlights challenges such as inadequate meaning representation, factual consistency, controllable text summarization, cross-lingual summarization, and evaluation metrics.
arXiv Detail & Related papers (2024-09-04T03:39:23Z) - Zero-Shot On-the-Fly Event Schema Induction [61.91468909200566]
We present a new approach in which large language models are utilized to generate source documents that allow predicting, given a high-level event definition, the specific events, arguments, and relations between them.
Using our model, complete schemas on any topic can be generated on-the-fly without any manual data collection, i.e., in a zero-shot manner.
arXiv Detail & Related papers (2022-10-12T14:37:00Z) - Embedding Knowledge for Document Summarization: A Survey [66.76415502727802]
Previous works proved that knowledge-embedded document summarizers excel at generating superior digests.
We propose novel to recapitulate knowledge and knowledge embeddings under the document summarization view.
arXiv Detail & Related papers (2022-04-24T04:36:07Z) - Aspect-Controllable Opinion Summarization [58.5308638148329]
We propose an approach that allows the generation of customized summaries based on aspect queries.
Using a review corpus, we create a synthetic training dataset of (review, summary) pairs enriched with aspect controllers.
We fine-tune a pretrained model using our synthetic dataset and generate aspect-specific summaries by modifying the aspect controllers.
arXiv Detail & Related papers (2021-09-07T16:09:17Z) - Meta-Transfer Learning for Low-Resource Abstractive Summarization [12.757403709439325]
Low-Resource Abstractive Summarization aims to leverage past experience to improve the performance with limited labeled examples of target corpus.
We conduct extensive experiments on various summarization corpora with different writing styles and forms.
The results demonstrate that our approach achieves the state-of-the-art on 6 corpora in low-resource scenarios, with only 0.7% of trainable parameters compared to previous work.
arXiv Detail & Related papers (2021-02-18T14:42:09Z) - Topic-Guided Abstractive Text Summarization: a Joint Learning Approach [19.623946402970933]
We introduce a new approach for abstractive text summarization, Topic-Guided Abstractive Summarization.
The idea is to incorporate neural topic modeling with a Transformer-based sequence-to-sequence (seq2seq) model in a joint learning framework.
arXiv Detail & Related papers (2020-10-20T14:45:25Z) - Unsupervised Reference-Free Summary Quality Evaluation via Contrastive
Learning [66.30909748400023]
We propose to evaluate the summary qualities without reference summaries by unsupervised contrastive learning.
Specifically, we design a new metric which covers both linguistic qualities and semantic informativeness based on BERT.
Experiments on Newsroom and CNN/Daily Mail demonstrate that our new evaluation method outperforms other metrics even without reference summaries.
arXiv Detail & Related papers (2020-10-05T05:04:14Z) - From Standard Summarization to New Tasks and Beyond: Summarization with
Manifold Information [77.89755281215079]
Text summarization is the research area aiming at creating a short and condensed version of the original document.
In real-world applications, most of the data is not in a plain text format.
This paper focuses on the survey of these new summarization tasks and approaches in the real-world application.
arXiv Detail & Related papers (2020-05-10T14:59:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.