Automated Detection of Doxing on Twitter
- URL: http://arxiv.org/abs/2202.00879v1
- Date: Wed, 2 Feb 2022 05:04:34 GMT
- Title: Automated Detection of Doxing on Twitter
- Authors: Younes Karimi, Anna Squicciarini, Shomir Wilson
- Abstract summary: Doxing refers to the practice of disclosing sensitive personal information about a person without their consent.
We propose and evaluate a set of approaches for automatically detecting second- and third-party disclosures on Twitter of sensitive private information.
- Score: 3.463438487417909
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Doxing refers to the practice of disclosing sensitive personal information
about a person without their consent. This form of cyberbullying is an
unpleasant and sometimes dangerous phenomenon for online social networks.
Although prior work exists on automated identification of other types of
cyberbullying, a need exists for methods capable of detecting doxing on Twitter
specifically. We propose and evaluate a set of approaches for automatically
detecting second- and third-party disclosures on Twitter of sensitive private
information, a subset of which constitutes doxing. We summarize our findings of
common intentions behind doxing episodes and compare nine different approaches
for automated detection based on string-matching and one-hot encoded
heuristics, as well as word and contextualized string embedding representations
of tweets. We identify an approach providing 96.86% accuracy and 97.37% recall
using contextualized string embeddings and conclude by discussing the
practicality of our proposed methods.
Related papers
- Enabling Contextual Soft Moderation on Social Media through Contrastive Textual Deviation [11.577310745082894]
We propose to incorporate stance detection into existing automated soft-moderation pipelines.
We show that our approach can reduce contextual false positives from 20% to 2.1%.
arXiv Detail & Related papers (2024-07-30T15:37:05Z) - RedactBuster: Entity Type Recognition from Redacted Documents [13.172863061928899]
We propose RedactBuster, the first deanonymization model using sentence context to perform Named Entity Recognition on reacted text.
We test RedactBuster against the most effective redaction technique and evaluate it using the publicly available Text Anonymization Benchmark (TAB)
Our results show accuracy values up to 0.985 regardless of the document nature or entity type.
arXiv Detail & Related papers (2024-04-19T16:42:44Z) - Verifying the Robustness of Automatic Credibility Assessment [79.08422736721764]
Text classification methods have been widely investigated as a way to detect content of low credibility.
In some cases insignificant changes in input text can mislead the models.
We introduce BODEGA: a benchmark for testing both victim models and attack methods on misinformation detection tasks.
arXiv Detail & Related papers (2023-03-14T16:11:47Z) - Fact-Saboteurs: A Taxonomy of Evidence Manipulation Attacks against
Fact-Verification Systems [80.3811072650087]
We show that it is possible to subtly modify claim-salient snippets in the evidence and generate diverse and claim-aligned evidence.
The attacks are also robust against post-hoc modifications of the claim.
These attacks can have harmful implications on the inspectable and human-in-the-loop usage scenarios.
arXiv Detail & Related papers (2022-09-07T13:39:24Z) - An Adversarial Attack Analysis on Malicious Advertisement URL Detection
Framework [22.259444589459513]
Malicious advertisement URLs pose a security risk since they are the source of cyber-attacks.
Existing malicious URL detection techniques are limited and to handle unseen features as well as generalize to test data.
In this study, we extract a novel set of lexical and web-scrapped features and employ machine learning technique to set up system for fraudulent advertisement URLs detection.
arXiv Detail & Related papers (2022-04-27T20:06:22Z) - Zero-Query Transfer Attacks on Context-Aware Object Detectors [95.18656036716972]
Adversarial attacks perturb images such that a deep neural network produces incorrect classification results.
A promising approach to defend against adversarial attacks on natural multi-object scenes is to impose a context-consistency check.
We present the first approach for generating context-consistent adversarial attacks that can evade the context-consistency check.
arXiv Detail & Related papers (2022-03-29T04:33:06Z) - Watch Those Words: Video Falsification Detection Using Word-Conditioned
Facial Motion [82.06128362686445]
We propose a multi-modal semantic forensic approach to handle both cheapfakes and visually persuasive deepfakes.
We leverage the idea of attribution to learn person-specific biometric patterns that distinguish a given speaker from others.
Unlike existing person-specific approaches, our method is also effective against attacks that focus on lip manipulation.
arXiv Detail & Related papers (2021-12-21T01:57:04Z) - Identification of Twitter Bots based on an Explainable ML Framework: the
US 2020 Elections Case Study [72.61531092316092]
This paper focuses on the design of a novel system for identifying Twitter bots based on labeled Twitter data.
Supervised machine learning (ML) framework is adopted using an Extreme Gradient Boosting (XGBoost) algorithm.
Our study also deploys Shapley Additive Explanations (SHAP) for explaining the ML model predictions.
arXiv Detail & Related papers (2021-12-08T14:12:24Z) - Cybersecurity Misinformation Detection on Social Media: Case Studies on
Phishing Reports and Zoom's Threats [1.2387676601792899]
We propose novel approaches for detecting misinformation about cybersecurity and privacy threats on social media.
We developed a framework for detecting inaccurate phishing claims on Twitter.
We also proposed another framework for detecting misinformation related to Zoom's security and privacy threats on multiple platforms.
arXiv Detail & Related papers (2021-10-23T20:45:24Z) - Backdoor Attack against Speaker Verification [86.43395230456339]
We show that it is possible to inject the hidden backdoor for infecting speaker verification models by poisoning the training data.
We also demonstrate that existing backdoor attacks cannot be directly adopted in attacking speaker verification.
arXiv Detail & Related papers (2020-10-22T11:10:08Z) - A Few Topical Tweets are Enough for Effective User-Level Stance
Detection [8.118808561953514]
We tackle stance detection for vocal Twitter users using two approaches.
In the first approach, we improve user-level stance detection by representing tweets using contextualized embeddings.
In the second approach, we expand the tweets of a given user using their Twitter timeline tweets, and then we perform unsupervised classification of the user.
arXiv Detail & Related papers (2020-04-07T15:35:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.