Large-Scale and Multi-Perspective Opinion Summarization with Diverse
Review Subsets
- URL: http://arxiv.org/abs/2310.13340v1
- Date: Fri, 20 Oct 2023 08:08:13 GMT
- Title: Large-Scale and Multi-Perspective Opinion Summarization with Diverse
Review Subsets
- Authors: Han Jiang, Rui Wang, Zhihua Wei, Yu Li, Xinpeng Wang
- Abstract summary: SUBSUMM is a supervised summarization framework for large-scale multi-perspective opinion summarization.
It generates pros, cons, and verdict summaries from hundreds of input reviews.
- Score: 23.515892409202344
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Opinion summarization is expected to digest larger review sets and provide
summaries from different perspectives. However, most existing solutions are
deficient in epitomizing extensive reviews and offering opinion summaries from
various angles due to the lack of designs for information selection. To this
end, we propose SUBSUMM, a supervised summarization framework for large-scale
multi-perspective opinion summarization. SUBSUMM consists of a review sampling
strategy set and a two-stage training scheme. The sampling strategies take
sentiment orientation and contrastive information value into consideration,
with which the review subsets from different perspectives and quality levels
can be selected. Subsequently, the summarizer is encouraged to learn from the
sub-optimal and optimal subsets successively in order to capitalize on the
massive input. Experimental results on AmaSum and Rotten Tomatoes datasets
demonstrate that SUBSUMM is adept at generating pros, cons, and verdict
summaries from hundreds of input reviews. Furthermore, our in-depth analysis
verifies that the advanced selection of review subsets and the two-stage
training scheme are vital to boosting the summarization performance.
Related papers
- Measuring "Why" in Recommender Systems: a Comprehensive Survey on the
Evaluation of Explainable Recommendation [87.82664566721917]
This survey is based on more than 100 papers from top-tier conferences like IJCAI, AAAI, TheWebConf, Recsys, UMAP, and IUI.
arXiv Detail & Related papers (2022-02-14T02:58:55Z) - Learning Opinion Summarizers by Selecting Informative Reviews [81.47506952645564]
We collect a large dataset of summaries paired with user reviews for over 31,000 products, enabling supervised training.
The content of many reviews is not reflected in the human-written summaries, and, thus, the summarizer trained on random review subsets hallucinates.
We formulate the task as jointly learning to select informative subsets of reviews and summarizing the opinions expressed in these subsets.
arXiv Detail & Related papers (2021-09-09T15:01:43Z) - Aspect-Controllable Opinion Summarization [58.5308638148329]
We propose an approach that allows the generation of customized summaries based on aspect queries.
Using a review corpus, we create a synthetic training dataset of (review, summary) pairs enriched with aspect controllers.
We fine-tune a pretrained model using our synthetic dataset and generate aspect-specific summaries by modifying the aspect controllers.
arXiv Detail & Related papers (2021-09-07T16:09:17Z) - Multi-Perspective Abstractive Answer Summarization [76.10437565615138]
Community Question Answering forums contain a rich resource of answers to a wide range of questions.
The goal of multi-perspective answer summarization is to produce a summary that includes all perspectives of the answer.
This work introduces a novel dataset creation method to automatically create multi-perspective, bullet-point abstractive summaries.
arXiv Detail & Related papers (2021-04-17T13:15:29Z) - Extractive Opinion Summarization in Quantized Transformer Spaces [52.95867345952894]
We present the Quantized Transformer (QT), an unsupervised system for extractive opinion summarization.
QT is inspired by Vector-Quantized Variational Autoencoders, which we repurpose for popularity-driven summarization.
arXiv Detail & Related papers (2020-12-08T14:23:46Z) - Unsupervised Reference-Free Summary Quality Evaluation via Contrastive
Learning [66.30909748400023]
We propose to evaluate the summary qualities without reference summaries by unsupervised contrastive learning.
Specifically, we design a new metric which covers both linguistic qualities and semantic informativeness based on BERT.
Experiments on Newsroom and CNN/Daily Mail demonstrate that our new evaluation method outperforms other metrics even without reference summaries.
arXiv Detail & Related papers (2020-10-05T05:04:14Z) - Massive Multi-Document Summarization of Product Reviews with Weak
Supervision [11.462916848094403]
Product reviews summarization is a type of Multi-Document Summarization (MDS) task.
We show that summarizing small samples of the reviews can result in loss of important information.
We propose a schema for summarizing a massive set of reviews on top of a standard summarization algorithm.
arXiv Detail & Related papers (2020-07-22T11:22:57Z) - Read what you need: Controllable Aspect-based Opinion Summarization of
Tourist Reviews [23.7107052882747]
We argue the need and propose a solution for generating personalized aspect-based opinion summaries from online tourist reviews.
We let our readers decide and control several attributes of the summary such as the length and specific aspects of interest.
Specifically, we take an unsupervised approach to extract coherent aspects from tourist reviews posted on TripAdvisor.
arXiv Detail & Related papers (2020-06-08T15:03:38Z) - OpinionDigest: A Simple Framework for Opinion Summarization [22.596995566588422]
The framework uses an Aspect-based Sentiment Analysis model to extract opinion phrases from reviews, and trains a Transformer model to reconstruct the original reviews from these extractions.
The selected opinions are used as input to the trained Transformer model, which verbalizes them into an opinion summary.
OpinionDigest can also generate customized summaries, tailored to specific user needs, by filtering the selected opinions according to their aspect and/or sentiment.
arXiv Detail & Related papers (2020-05-05T01:22:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.