Detecting and Reasoning of Deleted Tweets before they are Posted
- URL: http://arxiv.org/abs/2305.04927v1
- Date: Fri, 5 May 2023 08:25:07 GMT
- Title: Detecting and Reasoning of Deleted Tweets before they are Posted
- Authors: Hamdy Mubarak, Samir Abdaljalil, Azza Nassar and Firoj Alam
- Abstract summary: We identify deleted tweets, particularly within the Arabic context, and label them with a corresponding fine-grained disinformation category.
We then develop models that can predict the potentiality of tweets getting deleted, as well as the potential reasons behind deletion.
- Score: 5.300190188468289
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Social media platforms empower us in several ways, from information
dissemination to consumption. While these platforms are useful in promoting
citizen journalism, public awareness etc., they have misuse potentials.
Malicious users use them to disseminate hate-speech, offensive content, rumor
etc. to gain social and political agendas or to harm individuals, entities and
organizations. Often times, general users unconsciously share information
without verifying it, or unintentionally post harmful messages. Some of such
content often get deleted either by the platform due to the violation of terms
and policies, or users themselves for different reasons, e.g., regrets. There
is a wide range of studies in characterizing, understanding and predicting
deleted content. However, studies which aims to identify the fine-grained
reasons (e.g., posts are offensive, hate speech or no identifiable reason)
behind deleted content, are limited. In this study we address this gap, by
identifying deleted tweets, particularly within the Arabic context, and
labeling them with a corresponding fine-grained disinformation category. We
then develop models that can predict the potentiality of tweets getting
deleted, as well as the potential reasons behind deletion. Such models can help
in moderating social media posts before even posting.
Related papers
- Can Sensitive Information Be Deleted From LLMs? Objectives for Defending
Against Extraction Attacks [73.53327403684676]
We propose an attack-and-defense framework for studying the task of deleting sensitive information directly from model weights.
We study direct edits to model weights because this approach should guarantee that particular deleted information is never extracted by future prompt attacks.
We show that even state-of-the-art model editing methods such as ROME struggle to truly delete factual information from models like GPT-J, as our whitebox and blackbox attacks can recover "deleted" information from an edited model 38% of the time.
arXiv Detail & Related papers (2023-09-29T17:12:43Z) - Topological Data Mapping of Online Hate Speech, Misinformation, and
General Mental Health: A Large Language Model Based Study [6.803493330690884]
Recent advances in machine learning and large language models have made such an analysis possible.
In this study, we collected thousands of posts from carefully selected communities on the social media site Reddit.
We performed various machine-learning classifications based on embeddings in order to understand the role of hate speech/misinformation in various communities.
arXiv Detail & Related papers (2023-09-22T15:10:36Z) - Folk Models of Misinformation on Social Media [10.667165962654996]
We identify at least five folk models that conceptualize misinformation as either: political (counter)argumentation, out-of-context narratives, inherently fallacious information, external propaganda, or simply entertainment.
We use the rich conceptualizations embodied in these folk models to uncover how social media users minimize adverse reactions to misinformation encounters in their everyday lives.
arXiv Detail & Related papers (2022-07-26T00:40:26Z) - Manipulating Twitter Through Deletions [64.33261764633504]
Research into influence campaigns on Twitter has mostly relied on identifying malicious activities from tweets obtained via public APIs.
Here, we provide the first exhaustive, large-scale analysis of anomalous deletion patterns involving more than a billion deletions by over 11 million accounts.
We find that a small fraction of accounts delete a large number of tweets daily.
First, limits on tweet volume are circumvented, allowing certain accounts to flood the network with over 26 thousand daily tweets.
Second, coordinated networks of accounts engage in repetitive likes and unlikes of content that is eventually deleted, which can manipulate ranking algorithms.
arXiv Detail & Related papers (2022-03-25T20:07:08Z) - Beyond Plain Toxic: Detection of Inappropriate Statements on Flammable
Topics for the Russian Language [76.58220021791955]
We present two text collections labelled according to binary notion of inapropriateness and a multinomial notion of sensitive topic.
To objectivise the notion of inappropriateness, we define it in a data-driven way though crowdsourcing.
arXiv Detail & Related papers (2022-03-04T15:59:06Z) - News consumption and social media regulations policy [70.31753171707005]
We analyze two social media that enforced opposite moderation methods, Twitter and Gab, to assess the interplay between news consumption and content regulation.
Our results show that the presence of moderation pursued by Twitter produces a significant reduction of questionable content.
The lack of clear regulation on Gab results in the tendency of the user to engage with both types of content, showing a slight preference for the questionable ones which may account for a dissing/endorsement behavior.
arXiv Detail & Related papers (2021-06-07T19:26:32Z) - "I Won the Election!": An Empirical Analysis of Soft Moderation
Interventions on Twitter [0.9391375268580806]
We study the users who share tweets with warning labels on Twitter and their political leaning.
We find that 72% of the tweets with warning labels are shared by Republicans, while only 11% are shared by Democrats.
arXiv Detail & Related papers (2021-01-18T17:39:58Z) - Deceptive Deletions for Protecting Withdrawn Posts on Social Platforms [21.924023700334065]
We introduce Deceptive Deletion, a decoy mechanism that minimizes the adversarial advantage.
We show that a powerful global adversary can be beaten by a powerful challenger.
arXiv Detail & Related papers (2020-05-28T16:08:33Z) - Racism is a Virus: Anti-Asian Hate and Counterspeech in Social Media
during the COVID-19 Crisis [51.39895377836919]
COVID-19 has sparked racism and hate on social media targeted towards Asian communities.
We study the evolution and spread of anti-Asian hate speech through the lens of Twitter.
We create COVID-HATE, the largest dataset of anti-Asian hate and counterspeech spanning 14 months.
arXiv Detail & Related papers (2020-05-25T21:58:09Z) - Echo Chambers on Social Media: A comparative analysis [64.2256216637683]
We introduce an operational definition of echo chambers and perform a massive comparative analysis on 1B pieces of contents produced by 1M users on four social media platforms.
We infer the leaning of users about controversial topics and reconstruct their interaction networks by analyzing different features.
We find support for the hypothesis that platforms implementing news feed algorithms like Facebook may elicit the emergence of echo-chambers.
arXiv Detail & Related papers (2020-04-20T20:00:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.