PopALM: Popularity-Aligned Language Models for Social Media Trendy
Response Prediction
- URL: http://arxiv.org/abs/2402.18950v1
- Date: Thu, 29 Feb 2024 08:28:04 GMT
- Title: PopALM: Popularity-Aligned Language Models for Social Media Trendy
Response Prediction
- Authors: Erxin Yu, Jing Li, Chunpu Xu
- Abstract summary: We study trendy response prediction to automatically generate top-liked user replies to social media events.
We propose Popularity-Aligned Language Models (PopALM) to distinguish responses liked by a larger audience through reinforcement learning.
In experiments, we build a large-scale Weibo dataset for trendy response prediction, and its results show that PopALM can help boost the performance of advanced language models.
- Score: 6.979995957338177
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Social media platforms are daily exhibiting millions of events. To
preliminarily predict the mainstream public reaction to these events, we study
trendy response prediction to automatically generate top-liked user replies to
social media events. While previous works focus on generating responses without
factoring in popularity, we propose Popularity-Aligned Language Models (PopALM)
to distinguish responses liked by a larger audience through reinforcement
learning. Recognizing the noisy labels from user "likes", we tailor-make
curriculum learning in proximal policy optimization (PPO) to help models
capture the essential samples for easy-to-hard training. In experiments, we
build a large-scale Weibo dataset for trendy response prediction, and its
results show that PopALM can help boost the performance of advanced language
models.
Related papers
- Speechworthy Instruction-tuned Language Models [71.8586707840169]
We show that both prompting and preference learning increase the speech-suitability of popular instruction-tuned LLMs.
We share lexical, syntactical, and qualitative analyses to showcase how each method contributes to improving the speech-suitability of generated responses.
arXiv Detail & Related papers (2024-09-23T02:34:42Z) - Generative Deduplication For Socia Media Data Selection [4.545354973721937]
We propose a novel Generative Deduplication framework for social media data selection.
Our model acts as an efficient pre-processing method to universally enhance social media NLP pipelines.
arXiv Detail & Related papers (2024-01-11T12:43:26Z) - Decoding the Silent Majority: Inducing Belief Augmented Social Graph
with Large Language Model for Response Forecasting [74.68371461260946]
SocialSense is a framework that induces a belief-centered graph on top of an existent social network, along with graph-based propagation to capture social dynamics.
Our method surpasses existing state-of-the-art in experimental evaluations for both zero-shot and supervised settings.
arXiv Detail & Related papers (2023-10-20T06:17:02Z) - Organized Event Participant Prediction Enhanced by Social Media
Retweeting Data [8.675064911866201]
We propose to utilize social media retweeting activity data to enhance the learning of event participant prediction models.
We conduct comprehensive experiments in two scenarios with real-world data.
arXiv Detail & Related papers (2023-10-02T04:26:07Z) - Measuring the Effect of Influential Messages on Varying Personas [67.1149173905004]
We present a new task, Response Forecasting on Personas for News Media, to estimate the response a persona might have upon seeing a news message.
The proposed task not only introduces personalization in the modeling but also predicts the sentiment polarity and intensity of each response.
This enables more accurate and comprehensive inference on the mental state of the persona.
arXiv Detail & Related papers (2023-05-25T21:01:00Z) - SimOAP: Improve Coherence and Consistency in Persona-based Dialogue
Generation via Over-sampling and Post-evaluation [54.66399120084227]
Language models trained on large-scale corpora can generate remarkably fluent results in open-domain dialogue.
For the persona-based dialogue generation task, consistency and coherence are great challenges for language models.
A two-stage SimOAP strategy is proposed, i.e., over-sampling and post-evaluation.
arXiv Detail & Related papers (2023-05-18T17:23:00Z) - ASPEST: Bridging the Gap Between Active Learning and Selective
Prediction [56.001808843574395]
Selective prediction aims to learn a reliable model that abstains from making predictions when uncertain.
Active learning aims to lower the overall labeling effort, and hence human dependence, by querying the most informative examples.
In this work, we introduce a new learning paradigm, active selective prediction, which aims to query more informative samples from the shifted target domain.
arXiv Detail & Related papers (2023-04-07T23:51:07Z) - Towards Proactively Forecasting Sentence-Specific Information Popularity
within Online News Documents [13.537665342333488]
We introduce the task of proactively forecasting popularities of sentences within online news documents.
For training our models, we curate InfoPop, the first dataset containing popularity labels for over 1.7 million sentences.
We propose a novel transfer learning approach involving sentence salience prediction as an auxiliary task.
arXiv Detail & Related papers (2022-12-31T08:40:08Z) - TwHIN-BERT: A Socially-Enriched Pre-trained Language Model for
Multilingual Tweet Representations at Twitter [31.698196219228024]
We present TwHIN-BERT, a multilingual language model productionized at Twitter.
Our model is trained on 7 billion tweets covering over 100 distinct languages.
We evaluate our model on various multilingual social recommendation and semantic understanding tasks.
arXiv Detail & Related papers (2022-09-15T19:01:21Z) - Dialogue Response Ranking Training with Large-Scale Human Feedback Data [52.12342165926226]
We leverage social media feedback data to build a large-scale training dataset for feedback prediction.
We trained DialogRPT, a set of GPT-2 based models on 133M pairs of human feedback data.
Our ranker outperforms the conventional dialog perplexity baseline with a large margin on predicting Reddit feedback.
arXiv Detail & Related papers (2020-09-15T10:50:05Z) - On the Limits to Multi-Modal Popularity Prediction on Instagram -- A New
Robust, Efficient and Explainable Baseline [5.859055059050023]
We present a robust, efficient, and explainable baseline for population-based popularity prediction.
We employ the latest methods in computer vision to maximize the information extracted from the visual modality.
Our strongest models inform a lower limit to population-based predictability of popularity on Instagram.
arXiv Detail & Related papers (2020-04-26T21:21:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.