SocialPET: Socially Informed Pattern Exploiting Training for Few-Shot
Stance Detection in Social Media
- URL: http://arxiv.org/abs/2403.05216v1
- Date: Fri, 8 Mar 2024 11:00:09 GMT
- Title: SocialPET: Socially Informed Pattern Exploiting Training for Few-Shot
Stance Detection in Social Media
- Authors: Parisa Jamadi Khiabani, Arkaitz Zubiaga
- Abstract summary: Stance detection is the task of determining the viewpoint of a social media post towards a target as 'favor' or 'against'
SocialPET is a socially informed approach to leveraging language models for the task.
We prove the effectiveness of SocialPET on two stance datasets, Multi-target and P-Stance.
- Score: 8.556183465416156
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Stance detection, as the task of determining the viewpoint of a social media
post towards a target as 'favor' or 'against', has been understudied in the
challenging yet realistic scenario where there is limited labeled data for a
certain target. Our work advances research in few-shot stance detection by
introducing SocialPET, a socially informed approach to leveraging language
models for the task. Our proposed approach builds on the Pattern Exploiting
Training (PET) technique, which addresses classification tasks as cloze
questions through the use of language models. To enhance the approach with
social awareness, we exploit the social network structure surrounding social
media posts. We prove the effectiveness of SocialPET on two stance datasets,
Multi-target and P-Stance, outperforming competitive stance detection models as
well as the base model, PET, where the labeled instances for the target under
study is as few as 100. When we delve into the results, we observe that
SocialPET is comparatively strong in identifying instances of the `against'
class, where baseline models underperform.
Related papers
- SoMeLVLM: A Large Vision Language Model for Social Media Processing [78.47310657638567]
We introduce a Large Vision Language Model for Social Media Processing (SoMeLVLM)
SoMeLVLM is a cognitive framework equipped with five key capabilities including knowledge & comprehension, application, analysis, evaluation, and creation.
Our experiments demonstrate that SoMeLVLM achieves state-of-the-art performance in multiple social media tasks.
arXiv Detail & Related papers (2024-02-20T14:02:45Z) - Social-LLM: Modeling User Behavior at Scale using Language Models and
Social Network Data [13.660150473547766]
We introduce a novel approach tailored for modeling social network data in user detection tasks.
Our method integrates localized social network interactions with the capabilities of large language models.
We conduct a thorough evaluation of our method across seven real-world social network datasets.
arXiv Detail & Related papers (2023-12-31T05:13:13Z) - DeSIQ: Towards an Unbiased, Challenging Benchmark for Social
Intelligence Understanding [60.84356161106069]
We study the soundness of Social-IQ, a dataset of multiple-choice questions on videos of complex social interactions.
Our analysis reveals that Social-IQ contains substantial biases, which can be exploited by a moderately strong language model.
We introduce DeSIQ, a new challenging dataset, constructed by applying simple perturbations to Social-IQ.
arXiv Detail & Related papers (2023-10-24T06:21:34Z) - SOTOPIA: Interactive Evaluation for Social Intelligence in Language Agents [107.4138224020773]
We present SOTOPIA, an open-ended environment to simulate complex social interactions between artificial agents and humans.
In our environment, agents role-play and interact under a wide variety of scenarios; they coordinate, collaborate, exchange, and compete with each other to achieve complex social goals.
We find that GPT-4 achieves a significantly lower goal completion rate than humans and struggles to exhibit social commonsense reasoning and strategic communication skills.
arXiv Detail & Related papers (2023-10-18T02:27:01Z) - Few-shot Learning for Cross-Target Stance Detection by Aggregating
Multimodal Embeddings [16.39344929765961]
We introduce CT-TN, a novel model that aggregates multimodal embeddings from both textual and network features of the data.
We conduct experiments in a few-shot cross-target scenario on six different combinations of source-destination target pairs.
Experiments with different numbers of shots show that CT-TN can outperform other models after seeing 300 instances of the destination target.
arXiv Detail & Related papers (2023-01-11T15:52:55Z) - Improved Target-specific Stance Detection on Social Media Platforms by
Delving into Conversation Threads [12.007570049217398]
We propose a new task called conversational stance detection.
It infers the stance towards a given target (e.g., COVID-19 vaccination) when given a data instance and its corresponding conversation thread.
To infer the desired stances from both data instances and conversation threads, we propose a model called Branch-BERT that incorporates contextual information in conversation threads.
arXiv Detail & Related papers (2022-11-06T08:40:48Z) - Co-Located Human-Human Interaction Analysis using Nonverbal Cues: A
Survey [71.43956423427397]
We aim to identify the nonverbal cues and computational methodologies resulting in effective performance.
This survey differs from its counterparts by involving the widest spectrum of social phenomena and interaction settings.
Some major observations are: the most often used nonverbal cue, computational method, interaction environment, and sensing approach are speaking activity, support vector machines, and meetings composed of 3-4 persons equipped with microphones and cameras, respectively.
arXiv Detail & Related papers (2022-07-20T13:37:57Z) - Social Processes: Self-Supervised Forecasting of Nonverbal Cues in
Social Conversations [22.302509912465077]
We take the first step in the direction of a bottom-up self-supervised approach in the domain of social human interactions.
We formulate the task of Social Cue Forecasting to leverage the larger amount of unlabeled low-level behavior cues.
We propose the Social Process (SP) models--socially aware sequence-to-sequence (Seq2Seq) models within the Neural Process (NP) family.
arXiv Detail & Related papers (2021-07-28T18:01:08Z) - PHASE: PHysically-grounded Abstract Social Events for Machine Social
Perception [50.551003004553806]
We create a dataset of physically-grounded abstract social events, PHASE, that resemble a wide range of real-life social interactions.
Phase is validated with human experiments demonstrating that humans perceive rich interactions in the social events.
As a baseline model, we introduce a Bayesian inverse planning approach, SIMPLE, which outperforms state-of-the-art feed-forward neural networks.
arXiv Detail & Related papers (2021-03-02T18:44:57Z) - Human Trajectory Forecasting in Crowds: A Deep Learning Perspective [89.4600982169]
We present an in-depth analysis of existing deep learning-based methods for modelling social interactions.
We propose two knowledge-based data-driven methods to effectively capture these social interactions.
We develop a large scale interaction-centric benchmark TrajNet++, a significant yet missing component in the field of human trajectory forecasting.
arXiv Detail & Related papers (2020-07-07T17:19:56Z) - Stance Detection on Social Media: State of the Art and Trends [5.584060970507506]
Stance detection on social media is an emerging opinion mining paradigm for various social and political applications in which sentiment analysis may be sub-optimal.
This paper surveys the work on stance detection within those communities and situates its usage within current opinion mining techniques in social media.
It presents an exhaustive review of stance detection techniques on social media, including the task definition, different types of targets in stance detection, features set used, and various machine learning approaches applied.
arXiv Detail & Related papers (2020-06-05T19:24:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.