Proactive Prioritization of App Issues via Contrastive Learning
- URL: http://arxiv.org/abs/2303.06586v1
- Date: Sun, 12 Mar 2023 06:23:10 GMT
- Title: Proactive Prioritization of App Issues via Contrastive Learning
- Authors: Moghis Fereidouni, Adib Mosharrof, Umar Farooq, AB Siddique
- Abstract summary: We propose a new framework, PPrior, that enables proactive prioritization of app issues through identifying prominent reviews.
PPrior employs a pre-trained T5 model and works in three phases.
Phase one adapts the pre-trained T5 model to the user reviews data in a self-supervised fashion.
Phase two, we leverage contrastive training to learn a generic and task-independent representation of user reviews.
- Score: 2.6763498831034043
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Mobile app stores produce a tremendous amount of data in the form of user
reviews, which is a huge source of user requirements and sentiments; such
reviews allow app developers to proactively address issues in their apps.
However, only a small number of reviews capture common issues and sentiments
which creates a need for automatically identifying prominent reviews.
Unfortunately, most existing work in text ranking and popularity prediction
focuses on social contexts where other signals are available, which renders
such works ineffective in the context of app reviews. In this work, we propose
a new framework, PPrior, that enables proactive prioritization of app issues
through identifying prominent reviews (ones predicted to receive a large number
of votes in a given time window). Predicting highly-voted reviews is
challenging given that, unlike social posts, social network features of users
are not available. Moreover, there is an issue of class imbalance, since a
large number of user reviews receive little to no votes. PPrior employs a
pre-trained T5 model and works in three phases. Phase one adapts the
pre-trained T5 model to the user reviews data in a self-supervised fashion. In
phase two, we leverage contrastive training to learn a generic and
task-independent representation of user reviews. Phase three uses radius
neighbors classifier t o m ake t he final predictions. This phase also uses
FAISS index for scalability and efficient search. To conduct extensive
experiments, we acquired a large dataset of over 2.1 million user reviews from
Google Play. Our experimental results demonstrate the effectiveness of the
proposed framework when compared against several state-of-the-art approaches.
Moreover, the accuracy of PPrior in predicting prominent reviews is comparable
to that of experienced app developers.
Related papers
- Fairness Concerns in App Reviews: A Study on AI-based Mobile Apps [9.948068408730654]
This research aims to investigate fairness concerns raised in mobile app reviews.
Our research focuses on AI-based mobile app reviews as the chance of unfair behaviors and outcomes in AI-based apps may be higher than in non-AI-based apps.
arXiv Detail & Related papers (2024-01-16T03:43:33Z) - Inclusiveness Matters: A Large-Scale Analysis of User Feedback [7.8788463395442045]
We leverage user feedback from three popular online sources, Reddit, Google Play Store, and Twitter, for 50 of the most popular apps in the world.
Using a Socio-Technical Grounded Theory approach, we analyzed 23,107 posts across the three sources and identified 1,211 inclusiveness related posts.
Our study provides an in-depth view of inclusiveness-related user feedback from most popular apps and online sources.
arXiv Detail & Related papers (2023-11-02T04:05:46Z) - Decoding the Silent Majority: Inducing Belief Augmented Social Graph
with Large Language Model for Response Forecasting [74.68371461260946]
SocialSense is a framework that induces a belief-centered graph on top of an existent social network, along with graph-based propagation to capture social dynamics.
Our method surpasses existing state-of-the-art in experimental evaluations for both zero-shot and supervised settings.
arXiv Detail & Related papers (2023-10-20T06:17:02Z) - UltraFeedback: Boosting Language Models with Scaled AI Feedback [99.4633351133207]
We present textscUltraFeedback, a large-scale, high-quality, and diversified AI feedback dataset.
Our work validates the effectiveness of scaled AI feedback data in constructing strong open-source chat language models.
arXiv Detail & Related papers (2023-10-02T17:40:01Z) - Integrating Rankings into Quantized Scores in Peer Review [61.27794774537103]
In peer review, reviewers are usually asked to provide scores for the papers.
To mitigate this issue, conferences have started to ask reviewers to additionally provide a ranking of the papers they have reviewed.
There are no standard procedure for using this ranking information and Area Chairs may use it in different ways.
We take a principled approach to integrate the ranking information into the scores.
arXiv Detail & Related papers (2022-04-05T19:39:13Z) - SIFN: A Sentiment-aware Interactive Fusion Network for Review-based Item
Recommendation [48.1799451277808]
We propose a Sentiment-aware Interactive Fusion Network (SIFN) for review-based item recommendation.
We first encode user/item reviews via BERT and propose a light-weighted sentiment learner to extract semantic features of each review.
Then, we propose a sentiment prediction task that guides the sentiment learner to extract sentiment-aware features via explicit sentiment labels.
arXiv Detail & Related papers (2021-08-18T08:04:38Z) - Emerging App Issue Identification via Online Joint Sentiment-Topic
Tracing [66.57888248681303]
We propose a novel emerging issue detection approach named MERIT.
Based on the AOBST model, we infer the topics negatively reflected in user reviews for one app version.
Experiments on popular apps from Google Play and Apple's App Store demonstrate the effectiveness of MERIT.
arXiv Detail & Related papers (2020-08-23T06:34:05Z) - How Useful are Reviews for Recommendation? A Critical Review and
Potential Improvements [8.471274313213092]
We investigate a growing body of work that seeks to improve recommender systems through the use of review text.
Our initial findings reveal several discrepancies in reported results, partly due to copying results across papers despite changes in experimental settings or data pre-processing.
Further investigation calls for discussion on a much larger problem about the "importance" of user reviews for recommendation.
arXiv Detail & Related papers (2020-05-25T16:30:05Z) - Automating App Review Response Generation [67.58267006314415]
We propose a novel approach RRGen that automatically generates review responses by learning knowledge relations between reviews and their responses.
Experiments on 58 apps and 309,246 review-response pairs highlight that RRGen outperforms the baselines by at least 67.4% in terms of BLEU-4.
arXiv Detail & Related papers (2020-02-10T05:23:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.