HaT5: Hate Language Identification using Text-to-Text Transfer
Transformer
- URL: http://arxiv.org/abs/2202.05690v1
- Date: Fri, 11 Feb 2022 15:21:27 GMT
- Title: HaT5: Hate Language Identification using Text-to-Text Transfer
Transformer
- Authors: Sana Sabah Sabry, Tosin Adewumi, Nosheen Abid, Gy\"orgy Kovacs,
Foteini Liwicki and Marcus Liwicki
- Abstract summary: We investigate the performance of a state-of-the art (SoTA) architecture T5 across 5 different tasks from 2 relatively diverse datasets.
To improve performance, we augment the training data by using an autoregressive model.
It reveals the difficulties of poor data annotation by using a small set of examples.
- Score: 1.2532400738980594
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We investigate the performance of a state-of-the art (SoTA) architecture T5
(available on the SuperGLUE) and compare with it 3 other previous SoTA
architectures across 5 different tasks from 2 relatively diverse datasets. The
datasets are diverse in terms of the number and types of tasks they have. To
improve performance, we augment the training data by using an autoregressive
model. We achieve near-SoTA results on a couple of the tasks - macro F1 scores
of 81.66% for task A of the OLID 2019 dataset and 82.54% for task A of the hate
speech and offensive content (HASOC) 2021 dataset, where SoTA are 82.9% and
83.05%, respectively. We perform error analysis and explain why one of the
models (Bi-LSTM) makes the predictions it does by using a publicly available
algorithm: Integrated Gradient (IG). This is because explainable artificial
intelligence (XAI) is essential for earning the trust of users. The main
contributions of this work are the implementation method of T5, which is
discussed; the data augmentation using a new conversational AI model
checkpoint, which brought performance improvements; and the revelation on the
shortcomings of HASOC 2021 dataset. It reveals the difficulties of poor data
annotation by using a small set of examples where the T5 model made the correct
predictions, even when the ground truth of the test set were incorrect (in our
opinion). We also provide our model checkpoints on the HuggingFace hub1 to
foster transparency.
Related papers
- Data Interpreter: An LLM Agent For Data Science [43.13678782387546]
Large Language Model (LLM)-based agents have shown effectiveness across many applications.
However, their use in data science scenarios requiring solving long-term interconnected tasks, dynamic data adjustments and domain expertise remains challenging.
We present Data Interpreter, an LLM-based agent designed to automatically solve various data science problems end-to-end.
arXiv Detail & Related papers (2024-02-28T19:49:55Z) - Breaking Free Transformer Models: Task-specific Context Attribution
Promises Improved Generalizability Without Fine-tuning Pre-trained LLMs [1.5138606851862884]
We present a framework that allows for maintaining generalizability and enhances the performance on the downstream task.
We show that a linear transformation of the text representation from any transformer model using the task-specific concept operator results in a projection onto the latent concept space.
Experimental results on three datasets, namely HateXplain, IMDB reviews, and Social Media Attributions, illustrate that the proposed model attains superior accuracy and generalizability.
arXiv Detail & Related papers (2024-01-30T00:23:29Z) - Efficient Grammatical Error Correction Via Multi-Task Training and
Optimized Training Schedule [55.08778142798106]
We propose auxiliary tasks that exploit the alignment between the original and corrected sentences.
We formulate each task as a sequence-to-sequence problem and perform multi-task training.
We find that the order of datasets used for training and even individual instances within a dataset may have important effects on the final performance.
arXiv Detail & Related papers (2023-11-20T14:50:12Z) - T5 for Hate Speech, Augmented Data and Ensemble [1.3445335428144554]
We conduct investigations of automatic hate speech (HS) detection using different state-of-the-art (SoTA) baselines over 11 subtasks of 6 different datasets.
Our motivation is to determine which of the recent SoTA models is best for automatic hate speech detection and what advantage methods like data augmentation and ensemble may have on the best model, if any.
arXiv Detail & Related papers (2022-10-11T14:32:39Z) - Comparing Test Sets with Item Response Theory [53.755064720563]
We evaluate 29 datasets using predictions from 18 pretrained Transformer models on individual test examples.
We find that Quoref, HellaSwag, and MC-TACO are best suited for distinguishing among state-of-the-art models.
We also observe span selection task format, which is used for QA datasets like QAMR or SQuAD2.0, is effective in differentiating between strong and weak models.
arXiv Detail & Related papers (2021-06-01T22:33:53Z) - Learning to Perturb Word Embeddings for Out-of-distribution QA [55.103586220757464]
We propose a simple yet effective DA method based on a noise generator, which learns to perturb the word embedding of the input questions and context without changing their semantics.
We validate the performance of the QA models trained with our word embedding on a single source dataset, on five different target domains.
Notably, the model trained with ours outperforms the model trained with more than 240K artificially generated QA pairs.
arXiv Detail & Related papers (2021-05-06T14:12:26Z) - When Can Models Learn From Explanations? A Formal Framework for
Understanding the Roles of Explanation Data [84.87772675171412]
We study the circumstances under which explanations of individual data points can improve modeling performance.
We make use of three existing datasets with explanations: e-SNLI, TACRED, SemEval.
arXiv Detail & Related papers (2021-02-03T18:57:08Z) - Chatbot Interaction with Artificial Intelligence: Human Data
Augmentation with T5 and Language Transformer Ensemble for Text
Classification [2.492300648514128]
We present the Interaction with Artificial Intelligence (CI-AI) framework as an approach to the training of deep learning chatbots for task classification.
The intelligent system augments human-sourced data via artificial paraphrasing in order to generate a large set of training data.
We find that all models are improved when training data is augmented by the T5 model.
arXiv Detail & Related papers (2020-10-12T19:37:18Z) - TACRED Revisited: A Thorough Evaluation of the TACRED Relation
Extraction Task [80.38130122127882]
TACRED is one of the largest, most widely used crowdsourced datasets in Relation Extraction (RE)
In this paper, we investigate the questions: Have we reached a performance ceiling or is there still room for improvement?
We find that label errors account for 8% absolute F1 test error, and that more than 50% of the examples need to be relabeled.
arXiv Detail & Related papers (2020-04-30T15:07:37Z) - iTAML: An Incremental Task-Agnostic Meta-learning Approach [123.10294801296926]
Humans can continuously learn new knowledge as their experience grows.
Previous learning in deep neural networks can quickly fade out when they are trained on a new task.
We introduce a novel meta-learning approach that seeks to maintain an equilibrium between all encountered tasks.
arXiv Detail & Related papers (2020-03-25T21:42:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.