Trawling for Trolling: A Dataset
- URL: http://arxiv.org/abs/2008.00525v1
- Date: Sun, 2 Aug 2020 17:23:55 GMT
- Title: Trawling for Trolling: A Dataset
- Authors: Hitkul, Karmanya Aggarwal, Pakhi Bamdev, Debanjan Mahata, Rajiv Ratn
Shah and Ponnurangam Kumaraguru
- Abstract summary: We present a dataset that models trolling as a subcategory of offensive content.
The dataset has 12,490 samples, split across 5 classes; Normal, Profanity, Trolling, Derogatory and Hate Speech.
- Score: 56.1778095945542
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The ability to accurately detect and filter offensive content automatically
is important to ensure a rich and diverse digital discourse. Trolling is a type
of hurtful or offensive content that is prevalent in social media, but is
underrepresented in datasets for offensive content detection. In this work, we
present a dataset that models trolling as a subcategory of offensive content.
The dataset was created by collecting samples from well-known datasets and
reannotating them along precise definitions of different categories of
offensive content. The dataset has 12,490 samples, split across 5 classes;
Normal, Profanity, Trolling, Derogatory and Hate Speech. It encompasses content
from Twitter, Reddit and Wikipedia Talk Pages. Models trained on our dataset
show appreciable performance without any significant hyperparameter tuning and
can potentially learn meaningful linguistic information effectively. We find
that these models are sensitive to data ablation which suggests that the
dataset is largely devoid of spurious statistical artefacts that could
otherwise distract and confuse classification models.
Related papers
- Hate Speech Detection Using Cross-Platform Social Media Data In English and German Language [6.200058263544999]
This study focuses on detecting bilingual hate speech in YouTube comments.
We include factors such as content similarity, definition similarity, and common hate words to measure the impact of datasets on performance.
The best performance was obtained by combining datasets from YouTube comments, Twitter, and Gab with an F1-score of 0.74 and 0.68 for English and German YouTube comments.
arXiv Detail & Related papers (2024-10-02T10:22:53Z) - Sexism Detection on a Data Diet [14.899608305188002]
We show how we can leverage influence scores to estimate the importance of a data point while training a model.
We evaluate the model performance trained on data pruned with different pruning strategies on three out-of-domain datasets.
arXiv Detail & Related papers (2024-06-07T12:39:54Z) - THOS: A Benchmark Dataset for Targeted Hate and Offensive Speech [2.7061497863588126]
THOS is a dataset of 8.3k tweets manually labeled with fine-grained annotations about the target of the message.
We demonstrate that this dataset makes it feasible to train classifiers, based on Large Language Models, to perform classification at this level of granularity.
arXiv Detail & Related papers (2023-11-11T00:30:31Z) - Into the LAIONs Den: Investigating Hate in Multimodal Datasets [67.21783778038645]
This paper investigates the effect of scaling datasets on hateful content through a comparative audit of two datasets: LAION-400M and LAION-2B.
We found that hate content increased by nearly 12% with dataset scale, measured both qualitatively and quantitatively.
We also found that filtering dataset contents based on Not Safe For Work (NSFW) values calculated based on images alone does not exclude all the harmful content in alt-text.
arXiv Detail & Related papers (2023-11-06T19:00:05Z) - On Guiding Visual Attention with Language Specification [76.08326100891571]
We use high-level language specification as advice for constraining the classification evidence to task-relevant features, instead of distractors.
We show that supervising spatial attention in this way improves performance on classification tasks with biased and noisy data.
arXiv Detail & Related papers (2022-02-17T22:40:19Z) - Ruddit: Norms of Offensiveness for English Reddit Comments [35.83156813452207]
We create the first dataset of English language Reddit comments that has fine-grained, real-valued scores between -1 and 1.
We show that the method produces highly reliable offensiveness scores.
We evaluate the ability of widely-used neural models to predict offensiveness scores on this new dataset.
arXiv Detail & Related papers (2021-06-10T11:27:47Z) - Competency Problems: On Finding and Removing Artifacts in Language Data [50.09608320112584]
We argue that for complex language understanding tasks, all simple feature correlations are spurious.
We theoretically analyze the difficulty of creating data for competency problems when human bias is taken into account.
arXiv Detail & Related papers (2021-04-17T21:34:10Z) - Learning to Model and Ignore Dataset Bias with Mixed Capacity Ensembles [66.15398165275926]
We propose a method that can automatically detect and ignore dataset-specific patterns, which we call dataset biases.
Our method trains a lower capacity model in an ensemble with a higher capacity model.
We show improvement in all settings, including a 10 point gain on the visual question answering dataset.
arXiv Detail & Related papers (2020-11-07T22:20:03Z) - Hidden Footprints: Learning Contextual Walkability from 3D Human Trails [70.01257397390361]
Current datasets only tell you where people are, not where they could be.
We first augment the set of valid, labeled walkable regions by propagating person observations between images, utilizing 3D information to create what we call hidden footprints.
We devise a training strategy designed for such sparse labels, combining a class-balanced classification loss with a contextual adversarial loss.
arXiv Detail & Related papers (2020-08-19T23:19:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.