Topic Detection and Summarization of User Reviews
- URL: http://arxiv.org/abs/2006.00148v1
- Date: Sat, 30 May 2020 02:19:08 GMT
- Title: Topic Detection and Summarization of User Reviews
- Authors: Pengyuan Li, Lei Huang, Guang-jie Ren
- Abstract summary: We propose an effective new summarization method by analyzing both reviews and summaries.
A new dataset comprising product reviews and summaries about 1028 products are collected from Amazon and CNET.
- Score: 6.779855791259679
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: A massive amount of reviews are generated daily from various platforms. It is
impossible for people to read through tons of reviews and to obtain useful
information. Automatic summarizing customer reviews thus is important for
identifying and extracting the essential information to help users to obtain
the gist of the data. However, as customer reviews are typically short,
informal, and multifaceted, it is extremely challenging to generate topic-wise
summarization.While there are several studies aims to solve this issue, they
are heuristic methods that are developed only utilizing customer reviews.
Unlike existing method, we propose an effective new summarization method by
analyzing both reviews and summaries.To do that, we first segment reviews and
summaries into individual sentiments. As the sentiments are typically short, we
combine sentiments talking about the same aspect into a single document and
apply topic modeling method to identify hidden topics among customer reviews
and summaries. Sentiment analysis is employed to distinguish positive and
negative opinions among each detected topic. A classifier is also introduced to
distinguish the writing pattern of summaries and that of customer reviews.
Finally, sentiments are selected to generate the summarization based on their
topic relevance, sentiment analysis score and the writing pattern. To test our
method, a new dataset comprising product reviews and summaries about 1028
products are collected from Amazon and CNET. Experimental results show the
effectiveness of our method compared with other methods.
Related papers
- GLIMPSE: Pragmatically Informative Multi-Document Summarization for Scholarly Reviews [25.291384842659397]
We introduce sys, a summarization method designed to offer a concise yet comprehensive overview of scholarly reviews.
Unlike traditional consensus-based methods, sys extracts both common and unique opinions from the reviews.
arXiv Detail & Related papers (2024-06-11T15:27:01Z) - Incremental Extractive Opinion Summarization Using Cover Trees [81.59625423421355]
In online marketplaces user reviews accumulate over time, and opinion summaries need to be updated periodically.
In this work, we study the task of extractive opinion summarization in an incremental setting.
We present an efficient algorithm for accurately computing the CentroidRank summaries in an incremental setting.
arXiv Detail & Related papers (2024-01-16T02:00:17Z) - Towards Personalized Review Summarization by Modeling Historical Reviews
from Customer and Product Separately [59.61932899841944]
Review summarization is a non-trivial task that aims to summarize the main idea of the product review in the E-commerce website.
We propose the Heterogeneous Historical Review aware Review Summarization Model (HHRRS)
We employ a multi-task framework that conducts the review sentiment classification and summarization jointly.
arXiv Detail & Related papers (2023-01-27T12:32:55Z) - Learning Opinion Summarizers by Selecting Informative Reviews [81.47506952645564]
We collect a large dataset of summaries paired with user reviews for over 31,000 products, enabling supervised training.
The content of many reviews is not reflected in the human-written summaries, and, thus, the summarizer trained on random review subsets hallucinates.
We formulate the task as jointly learning to select informative subsets of reviews and summarizing the opinions expressed in these subsets.
arXiv Detail & Related papers (2021-09-09T15:01:43Z) - Polarity in the Classroom: A Case Study Leveraging Peer Sentiment Toward
Scalable Assessment [4.588028371034406]
Accurately grading open-ended assignments in large or massive open online courses (MOOCs) is non-trivial.
In this work, we detail the process by which we create our domain-dependent lexicon and aspect-informed review form.
We end by analyzing validity and discussing conclusions from our corpus of over 6800 peer reviews from nine courses.
arXiv Detail & Related papers (2021-08-02T15:45:11Z) - A Unified Dual-view Model for Review Summarization and Sentiment
Classification with Inconsistency Loss [51.448615489097236]
Acquiring accurate summarization and sentiment from user reviews is an essential component of modern e-commerce platforms.
We propose a novel dual-view model that jointly improves the performance of these two tasks.
Experiment results on four real-world datasets from different domains demonstrate the effectiveness of our model.
arXiv Detail & Related papers (2020-06-02T13:34:11Z) - Few-Shot Learning for Opinion Summarization [117.70510762845338]
Opinion summarization is the automatic creation of text reflecting subjective information expressed in multiple documents.
In this work, we show that even a handful of summaries is sufficient to bootstrap generation of the summary text.
Our approach substantially outperforms previous extractive and abstractive methods in automatic and human evaluation.
arXiv Detail & Related papers (2020-04-30T15:37:38Z) - Unsupervised Opinion Summarization with Noising and Denoising [85.49169453434554]
We create a synthetic dataset from a corpus of user reviews by sampling a review, pretending it is a summary, and generating noisy versions thereof.
At test time, the model accepts genuine reviews and generates a summary containing salient opinions, treating those that do not reach consensus as noise.
arXiv Detail & Related papers (2020-04-21T16:54:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.