Detecting and Characterizing Extremist Reviewer Groups in Online Product
Reviews
- URL: http://arxiv.org/abs/2004.05865v1
- Date: Mon, 13 Apr 2020 10:59:21 GMT
- Title: Detecting and Characterizing Extremist Reviewer Groups in Online Product
Reviews
- Authors: Viresh Gupta, Aayush Aggarwal, Tanmoy Chakraborty
- Abstract summary: In this paper, we collect reviews from the Amazon product review site and manually labelled a set of 923 candidate reviewer groups.
The groups are extracted using frequent itemset mining over brand similarities such that users are clustered together if they have mutually reviewed (products of) a lot of brands.
We run multiple classifiers for the task of classifying a group based on the reviews written by the users of that group, to determine if the group shows signs of extremity.
- Score: 26.473021051027537
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Online marketplaces often witness opinion spam in the form of reviews. People
are often hired to target specific brands for promoting or impeding them by
writing highly positive or negative reviews. This often is done collectively in
groups. Although some previous studies attempted to identify and analyze such
opinion spam groups, little has been explored to spot those groups who target a
brand as a whole, instead of just products.
In this paper, we collected reviews from the Amazon product review site and
manually labelled a set of 923 candidate reviewer groups. The groups are
extracted using frequent itemset mining over brand similarities such that users
are clustered together if they have mutually reviewed (products of) a lot of
brands. We hypothesize that the nature of the reviewer groups is dependent on 8
features specific to a (group, brand) pair. We develop a feature-based
supervised model to classify candidate groups as extremist entities. We run
multiple classifiers for the task of classifying a group based on the reviews
written by the users of that group, to determine if the group shows signs of
extremity. A 3-layer Perceptron based classifier turns out to be the best
classifier. We further study the behaviours of such groups in detail to
understand the dynamics of brand-level opinion fraud better. These behaviours
include consistency in ratings, review sentiment, verified purchase, review
dates and helpful votes received on reviews. Surprisingly, we observe that
there are a lot of verified reviewers showing extreme sentiment, which on
further investigation leads to ways to circumvent existing mechanisms in place
to prevent unofficial incentives on Amazon.
Related papers
- Analytical and Empirical Study of Herding Effects in Recommendation Systems [72.6693986712978]
We study how to manage product ratings via rating aggregation rules and shortlisted representative reviews.
We show that proper recency aware rating aggregation rules can improve the speed of convergence in Amazon and TripAdvisor.
arXiv Detail & Related papers (2024-08-20T14:29:23Z) - Combat AI With AI: Counteract Machine-Generated Fake Restaurant Reviews
on Social Media [77.34726150561087]
We propose to leverage the high-quality elite Yelp reviews to generate fake reviews from the OpenAI GPT review creator.
We apply the model to predict non-elite reviews and identify the patterns across several dimensions.
We show that social media platforms are continuously challenged by machine-generated fake reviews.
arXiv Detail & Related papers (2023-02-10T19:40:10Z) - Detecting Spam Reviews on Vietnamese E-commerce Websites [0.0]
We propose the dataset called ViSpamReviews, which has a strict annotation procedure for detecting spam reviews on e-commerce platforms.
Our dataset consists of two tasks: the binary classification task for detecting whether a review is a spam or not and the multi-class classification task for identifying the type of spam.
The PhoBERT obtained the highest results on both tasks, 88.93% and 72.17%, respectively, by macro average F1 score.
arXiv Detail & Related papers (2022-07-27T10:37:14Z) - Are Your Reviewers Being Treated Equally? Discovering Subgroup
Structures to Improve Fairness in Spam Detection [13.26226951002133]
This paper addresses the challenges of defining, approximating, and utilizing a new subgroup structure for fair spam detection.
We first identify subgroup structures in the review graph that lead to discrepant accuracy in the groups.
Comprehensive comparisons against baselines on three large Yelp review datasets demonstrate that the subgroup membership can be identified and exploited for group fairness.
arXiv Detail & Related papers (2022-04-24T02:19:22Z) - Spatio-Temporal Graph Representation Learning for Fraudster Group
Detection [50.779498955162644]
Companies may hire fraudster groups to write fake reviews to either demote competitors or promote their own businesses.
To detect such groups, a common model is to represent fraudster groups' static networks.
We propose to first capitalize on the effectiveness of the HIN-RNN in both reviewers' representation learning.
arXiv Detail & Related papers (2022-01-07T08:01:38Z) - Attention-based Bidirectional LSTM for Deceptive Opinion Spam
Classification [0.0]
Fraudulent reviews are deliberately posted on various online review platforms to trick customers to buy, visit or distract against a product or a restaurant.
The work aims at detecting and classifying the reviews as deceptive or truthful.
It involves use of various deep learning techniques for classifying the reviews and an overview of proposed approach.
arXiv Detail & Related papers (2021-12-29T19:02:04Z) - Learning Opinion Summarizers by Selecting Informative Reviews [81.47506952645564]
We collect a large dataset of summaries paired with user reviews for over 31,000 products, enabling supervised training.
The content of many reviews is not reflected in the human-written summaries, and, thus, the summarizer trained on random review subsets hallucinates.
We formulate the task as jointly learning to select informative subsets of reviews and summarizing the opinions expressed in these subsets.
arXiv Detail & Related papers (2021-09-09T15:01:43Z) - Identifying Hijacked Reviews [21.600163117735857]
Review Hijacking is a new review manipulation tactic in which unethical sellers "hijack" an existing product page.
We propose a framework to generate synthetically labeled data for review hijacking by swapping products and reviews.
We then evaluate the potential of both a Twin LSTM network and BERT sequence pair to distinguish legitimate reviews from hijacked ones using this data.
arXiv Detail & Related papers (2021-07-07T20:43:36Z) - Abstractive Opinion Tagging [65.47649273721679]
In e-commerce, opinion tags refer to a ranked list of tags provided by the e-commerce platform that reflect characteristics of reviews of an item.
Current mechanisms for generating opinion tags rely on either manual or labelling methods, which is time-consuming and ineffective.
We propose an abstractive opinion tagging framework, named AOT-Net, to generate a ranked list of opinion tags given a large number of reviews.
arXiv Detail & Related papers (2021-01-18T05:08:15Z) - A Unified Dual-view Model for Review Summarization and Sentiment
Classification with Inconsistency Loss [51.448615489097236]
Acquiring accurate summarization and sentiment from user reviews is an essential component of modern e-commerce platforms.
We propose a novel dual-view model that jointly improves the performance of these two tasks.
Experiment results on four real-world datasets from different domains demonstrate the effectiveness of our model.
arXiv Detail & Related papers (2020-06-02T13:34:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.