Time-aware topic identification in social media with pre-trained
language models: A case study of electric vehicles
- URL: http://arxiv.org/abs/2210.05143v1
- Date: Tue, 11 Oct 2022 04:50:10 GMT
- Title: Time-aware topic identification in social media with pre-trained
language models: A case study of electric vehicles
- Authors: Byeongki Jeong, Janghyeok Yoon, Jaewoong Choi
- Abstract summary: We propose a time-aware topic identification approach with pre-trained language models.
The proposed approach consists of two stages: the dynamics-focused function for tracking time-varying topics with language models and the emergence-scoring function to examine future promising topics.
- Score: 1.2891210250935146
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recent extensively competitive business environment makes companies to keep
their eyes on social media, as there is a growing recognition over customer
languages (e.g., needs, interests, and complaints) as source of future
opportunities. This research avenue analysing social media data has received
much attention in academia, but their utilities are limited as most of methods
provide retrospective results. Moreover, the increasing number of
customer-generated contents and rapidly varying topics have made the necessity
of time-aware topic evolution analyses. Recently, several researchers have
showed the applicability of pre-trained semantic language models to social
media as an input feature, but leaving limitations in understanding evolving
topics. In this study, we propose a time-aware topic identification approach
with pre-trained language models. The proposed approach consists of two stages:
the dynamics-focused function for tracking time-varying topics with language
models and the emergence-scoring function to examine future promising topics.
Here we apply the proposed approach to reddit data on electric vehicles, and
our findings highlight the feasibility of capturing emerging customer topics
from voluminous social media in a time-aware manner.
Related papers
- The Imperative of Conversation Analysis in the Era of LLMs: A Survey of Tasks, Techniques, and Trends [64.99423243200296]
Conversation Analysis (CA) strives to uncover and analyze critical information from conversation data.
In this paper, we perform a thorough review and systematize CA task to summarize the existing related work.
We derive four key steps of CA from conversation scene reconstruction, to in-depth attribution analysis, and then to performing targeted training, finally generating conversations.
arXiv Detail & Related papers (2024-09-21T16:52:43Z) - Recent Advances in Hate Speech Moderation: Multimodality and the Role of Large Models [52.24001776263608]
This comprehensive survey delves into the recent strides in HS moderation.
We highlight the burgeoning role of large language models (LLMs) and large multimodal models (LMMs)
We identify existing gaps in research, particularly in the context of underrepresented languages and cultures.
arXiv Detail & Related papers (2024-01-30T03:51:44Z) - AutoConv: Automatically Generating Information-seeking Conversations
with Large Language Models [74.10293412011455]
We propose AutoConv for synthetic conversation generation.
Specifically, we formulate the conversation generation problem as a language modeling task.
We finetune an LLM with a few human conversations to capture the characteristics of the information-seeking process.
arXiv Detail & Related papers (2023-08-12T08:52:40Z) - Lost in the Middle: How Language Models Use Long Contexts [88.78803442320246]
We analyze the performance of language models on two tasks that require identifying relevant information in their input contexts.
We find that performance can degrade significantly when changing the position of relevant information.
Our analysis provides a better understanding of how language models use their input context and provides new evaluation protocols for future long-context language models.
arXiv Detail & Related papers (2023-07-06T17:54:11Z) - Recent Advances in Direct Speech-to-text Translation [58.692782919570845]
We categorize the existing research work into three directions based on the main challenges -- modeling burden, data scarcity, and application issues.
For the challenge of data scarcity, recent work resorts to many sophisticated techniques, such as data augmentation, pre-training, knowledge distillation, and multilingual modeling.
We analyze and summarize the application issues, which include real-time, segmentation, named entity, gender bias, and code-switching.
arXiv Detail & Related papers (2023-06-20T16:14:27Z) - Few Shot Semantic Segmentation: a review of methodologies, benchmarks, and open challenges [5.0243930429558885]
Few-Shot Semantic is a novel task in computer vision, which aims at designing models capable of segmenting new semantic classes with only a few examples.
This paper consists of a comprehensive survey of Few-Shot Semantic, tracing its evolution and exploring various model designs.
arXiv Detail & Related papers (2023-04-12T13:07:37Z) - Analyzing social media with crowdsourcing in Crowd4SDG [1.1403672224109254]
This study presents an approach that provides flexible support for analyzing social media, particularly during emergencies.
The focus is on analyzing images and text contained in social media posts and a set of automatic data processing tools for filtering, classification, and geolocation of content.
Such support includes both feedback and suggestions to configure automated tools, and crowdsourcing to gather inputs from citizens.
arXiv Detail & Related papers (2022-08-04T14:42:20Z) - SocialVisTUM: An Interactive Visualization Toolkit for Correlated Neural
Topic Models on Social Media Opinion Mining [0.07538606213726905]
Recent research in opinion mining proposed word embedding-based topic modeling methods.
We show how these methods can be used to display correlated topic models on social media texts using SocialVisTUM.
arXiv Detail & Related papers (2021-10-20T14:04:13Z) - An NLP approach to quantify dynamic salience of predefined topics in a
text corpus [0.0]
We use natural language processing techniques to quantify how a set of pre-defined topics of interest change over time across a large corpus of text.
We find that given a predefined topic, we can identify and rank sets of terms, or n-grams, that map to those topics and have usage patterns that deviate from a normal baseline.
arXiv Detail & Related papers (2021-08-16T21:00:06Z) - Can You be More Social? Injecting Politeness and Positivity into
Task-Oriented Conversational Agents [60.27066549589362]
Social language used by human agents is associated with greater users' responsiveness and task completion.
The model uses a sequence-to-sequence deep learning architecture, extended with a social language understanding element.
Evaluation in terms of content preservation and social language level using both human judgment and automatic linguistic measures shows that the model can generate responses that enable agents to address users' issues in a more socially appropriate way.
arXiv Detail & Related papers (2020-12-29T08:22:48Z) - Keyword Assisted Topic Models [0.0]
We show that providing a small number of keywords can substantially enhance the measurement performance of topic models.
KeyATM provides more interpretable results, has better document classification performance, and is less sensitive to the number of topics than the standard topic models.
arXiv Detail & Related papers (2020-04-13T14:35:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.