Identifying Hijacked Reviews
- URL: http://arxiv.org/abs/2107.05385v1
- Date: Wed, 7 Jul 2021 20:43:36 GMT
- Title: Identifying Hijacked Reviews
- Authors: Monika Daryani and James Caverlee
- Abstract summary: Review Hijacking is a new review manipulation tactic in which unethical sellers "hijack" an existing product page.
We propose a framework to generate synthetically labeled data for review hijacking by swapping products and reviews.
We then evaluate the potential of both a Twin LSTM network and BERT sequence pair to distinguish legitimate reviews from hijacked ones using this data.
- Score: 21.600163117735857
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Fake reviews and review manipulation are growing problems on online
marketplaces globally. Review Hijacking is a new review manipulation tactic in
which unethical sellers "hijack" an existing product page (usually one with
many positive reviews), then update the product details like title, photo, and
description with those of an entirely different product. With the earlier
reviews still attached, the new item appears well-reviewed. However, there are
no public datasets of review hijacking and little is known in the literature
about this tactic. Hence, this paper proposes a three-part study: (i) we
propose a framework to generate synthetically labeled data for review hijacking
by swapping products and reviews; (ii) then, we evaluate the potential of both
a Twin LSTM network and BERT sequence pair classifier to distinguish legitimate
reviews from hijacked ones using this data; and (iii) we then deploy the best
performing model on a collection of 31K products (with 6.5 M reviews) in the
original data, where we find 100s of previously unknown examples of review
hijacking.
Related papers
- What Matters in Explanations: Towards Explainable Fake Review Detection Focusing on Transformers [45.55363754551388]
Customers' reviews and feedback play crucial role on e-commerce platforms like Amazon, Zalando, and eBay.
There is a prevailing concern that sellers often post fake or spam reviews to deceive potential customers and manipulate their opinions about a product.
We propose an explainable framework for detecting fake reviews with high precision in identifying fraudulent content with explanations.
arXiv Detail & Related papers (2024-07-24T13:26:02Z) - Enhanced Review Detection and Recognition: A Platform-Agnostic Approach with Application to Online Commerce [0.46040036610482665]
We present a machine learning methodology for review detection and extraction.
We demonstrate that it generalises for use across websites that were not contained in the training data.
This method promises to drive applications for automatic detection and evaluation of reviews, regardless of their source.
arXiv Detail & Related papers (2024-05-09T00:32:22Z) - Combat AI With AI: Counteract Machine-Generated Fake Restaurant Reviews
on Social Media [77.34726150561087]
We propose to leverage the high-quality elite Yelp reviews to generate fake reviews from the OpenAI GPT review creator.
We apply the model to predict non-elite reviews and identify the patterns across several dimensions.
We show that social media platforms are continuously challenged by machine-generated fake reviews.
arXiv Detail & Related papers (2023-02-10T19:40:10Z) - Towards Personalized Review Summarization by Modeling Historical Reviews
from Customer and Product Separately [59.61932899841944]
Review summarization is a non-trivial task that aims to summarize the main idea of the product review in the E-commerce website.
We propose the Heterogeneous Historical Review aware Review Summarization Model (HHRRS)
We employ a multi-task framework that conducts the review sentiment classification and summarization jointly.
arXiv Detail & Related papers (2023-01-27T12:32:55Z) - Detecting Spam Reviews on Vietnamese E-commerce Websites [0.0]
We propose the dataset called ViSpamReviews, which has a strict annotation procedure for detecting spam reviews on e-commerce platforms.
Our dataset consists of two tasks: the binary classification task for detecting whether a review is a spam or not and the multi-class classification task for identifying the type of spam.
The PhoBERT obtained the highest results on both tasks, 88.93% and 72.17%, respectively, by macro average F1 score.
arXiv Detail & Related papers (2022-07-27T10:37:14Z) - Attention-based Bidirectional LSTM for Deceptive Opinion Spam
Classification [0.0]
Fraudulent reviews are deliberately posted on various online review platforms to trick customers to buy, visit or distract against a product or a restaurant.
The work aims at detecting and classifying the reviews as deceptive or truthful.
It involves use of various deep learning techniques for classifying the reviews and an overview of proposed approach.
arXiv Detail & Related papers (2021-12-29T19:02:04Z) - Can We Automate Scientific Reviewing? [89.50052670307434]
We discuss the possibility of using state-of-the-art natural language processing (NLP) models to generate first-pass peer reviews for scientific papers.
We collect a dataset of papers in the machine learning domain, annotate them with different aspects of content covered in each review, and train targeted summarization models that take in papers to generate reviews.
Comprehensive experimental results show that system-generated reviews tend to touch upon more aspects of the paper than human-written reviews.
arXiv Detail & Related papers (2021-01-30T07:16:53Z) - DFraud3- Multi-Component Fraud Detection freeof Cold-start [50.779498955162644]
The Cold-start is a significant problem referring to the failure of a detection system to recognize the authenticity of a new user.
In this paper, we model a review system as a Heterogeneous InformationNetwork (HIN) which enables a unique representation to every component.
HIN with graph induction helps to address the camouflage issue (fraudsterswith genuine reviews) which has shown to be more severe when it is coupled with cold-start, i.e., new fraudsters with genuine first reviews.
arXiv Detail & Related papers (2020-06-10T08:20:13Z) - Context-aware Helpfulness Prediction for Online Product Reviews [34.47368084659301]
We propose a neural deep learning model that predicts the helpfulness score of a review.
This model is based on convolutional neural network (CNN) and a context-aware encoding mechanism.
We validated our model on human annotated dataset and the result shows that our model significantly outperforms existing models for helpfulness prediction.
arXiv Detail & Related papers (2020-04-27T18:19:26Z) - Unsupervised Opinion Summarization with Noising and Denoising [85.49169453434554]
We create a synthetic dataset from a corpus of user reviews by sampling a review, pretending it is a summary, and generating noisy versions thereof.
At test time, the model accepts genuine reviews and generates a summary containing salient opinions, treating those that do not reach consensus as noise.
arXiv Detail & Related papers (2020-04-21T16:54:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.