T5 for Hate Speech, Augmented Data and Ensemble
- URL: http://arxiv.org/abs/2210.05480v1
- Date: Tue, 11 Oct 2022 14:32:39 GMT
- Title: T5 for Hate Speech, Augmented Data and Ensemble
- Authors: Tosin Adewumi, Sana Sabah Sabry, Nosheen Abid, Foteini Liwicki and
Marcus Liwicki
- Abstract summary: We conduct investigations of automatic hate speech (HS) detection using different state-of-the-art (SoTA) baselines over 11 subtasks of 6 different datasets.
Our motivation is to determine which of the recent SoTA models is best for automatic hate speech detection and what advantage methods like data augmentation and ensemble may have on the best model, if any.
- Score: 1.3445335428144554
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We conduct relatively extensive investigations of automatic hate speech (HS)
detection using different state-of-the-art (SoTA) baselines over 11 subtasks of
6 different datasets. Our motivation is to determine which of the recent SoTA
models is best for automatic hate speech detection and what advantage methods
like data augmentation and ensemble may have on the best model, if any. We
carry out 6 cross-task investigations. We achieve new SoTA on two subtasks -
macro F1 scores of 91.73% and 53.21% for subtasks A and B of the HASOC 2020
dataset, where previous SoTA are 51.52% and 26.52%, respectively. We achieve
near-SoTA on two others - macro F1 scores of 81.66% for subtask A of the OLID
2019 dataset and 82.54% for subtask A of the HASOC 2021 dataset, where SoTA are
82.9% and 83.05%, respectively. We perform error analysis and use two
explainable artificial intelligence (XAI) algorithms (IG and SHAP) to reveal
how two of the models (Bi-LSTM and T5) make the predictions they do by using
examples. Other contributions of this work are 1) the introduction of a simple,
novel mechanism for correcting out-of-class (OOC) predictions in T5, 2) a
detailed description of the data augmentation methods, 3) the revelation of the
poor data annotations in the HASOC 2021 dataset by using several examples and
XAI (buttressing the need for better quality control), and 4) the public
release of our model checkpoints and codes to foster transparency.
Related papers
- Step-DPO: Step-wise Preference Optimization for Long-chain Reasoning of LLMs [54.05511925104712]
We propose a simple, effective, and data-efficient method called Step-DPO.
Step-DPO treats individual reasoning steps as units for preference optimization rather than evaluating answers holistically.
Our findings demonstrate that as few as 10K preference data pairs and fewer than 500 Step-DPO training steps can yield a nearly 3% gain in accuracy on MATH for models with over 70B parameters.
arXiv Detail & Related papers (2024-06-26T17:43:06Z) - Less is More: Fewer Interpretable Region via Submodular Subset Selection [54.07758302264416]
This paper re-models the above image attribution problem as a submodular subset selection problem.
We construct a novel submodular function to discover more accurate small interpretation regions.
For correctly predicted samples, the proposed method improves the Deletion and Insertion scores with an average of 4.9% and 2.5% gain relative to HSIC-Attribution.
arXiv Detail & Related papers (2024-02-14T13:30:02Z) - UU-Tax at SemEval-2022 Task 3: Improving the generalizability of
language models for taxonomy classification through data augmentation [0.0]
This paper addresses the SemEval-2022 Task 3 PreTENS: Presupposed Taxonomies evaluating Neural Network Semantics.
The goal of the task is to identify if a sentence is deemed acceptable or not, depending on the taxonomic relationship that holds between a noun pair contained in the sentence.
We propose an effective way to enhance the robustness and the generalizability of language models for better classification.
arXiv Detail & Related papers (2022-10-07T07:41:28Z) - HaT5: Hate Language Identification using Text-to-Text Transfer
Transformer [1.2532400738980594]
We investigate the performance of a state-of-the art (SoTA) architecture T5 across 5 different tasks from 2 relatively diverse datasets.
To improve performance, we augment the training data by using an autoregressive model.
It reveals the difficulties of poor data annotation by using a small set of examples.
arXiv Detail & Related papers (2022-02-11T15:21:27Z) - Reference-based Weak Supervision for Answer Sentence Selection using Web
Data [87.18646699292293]
We introduce Reference-based Weak Supervision (RWS), a fully automatic large-scale data pipeline.
RWS harvests high-quality weakly-supervised answers from abundant Web data.
Our experiments indicate that the produced data consistently bolsters TANDA.
arXiv Detail & Related papers (2021-04-18T19:41:17Z) - BreakingBERT@IITK at SemEval-2021 Task 9 : Statement Verification and
Evidence Finding with Tables [1.78256232654567]
We tackle the problem of fact verification and evidence finding over tabular data.
We make a comparison of the baselines and state-of-the-art approaches over the given SemTabFact dataset.
We also propose a novel approach CellBERT to solve evidence finding as a form of the Natural Language Inference task.
arXiv Detail & Related papers (2021-04-07T11:41:07Z) - LRG at SemEval-2021 Task 4: Improving Reading Comprehension with
Abstract Words using Augmentation, Linguistic Features and Voting [0.6850683267295249]
Given a fill-in-the-blank-type question, the task is to predict the most suitable word from a list of 5 options.
We use encoders of transformers-based models pre-trained on the masked language modelling (MLM) task to build our Fill-in-the-blank (FitB) models.
We propose variants, namely Chunk Voting and Max Context, to take care of input length restrictions for BERT, etc.
arXiv Detail & Related papers (2021-02-24T12:33:12Z) - Device-Robust Acoustic Scene Classification Based on Two-Stage
Categorization and Data Augmentation [63.98724740606457]
We present a joint effort of four groups, namely GT, USTC, Tencent, and UKE, to tackle Task 1 - Acoustic Scene Classification (ASC) in the DCASE 2020 Challenge.
Task 1a focuses on ASC of audio signals recorded with multiple (real and simulated) devices into ten different fine-grained classes.
Task 1b concerns with classification of data into three higher-level classes using low-complexity solutions.
arXiv Detail & Related papers (2020-07-16T15:07:14Z) - TACRED Revisited: A Thorough Evaluation of the TACRED Relation
Extraction Task [80.38130122127882]
TACRED is one of the largest, most widely used crowdsourced datasets in Relation Extraction (RE)
In this paper, we investigate the questions: Have we reached a performance ceiling or is there still room for improvement?
We find that label errors account for 8% absolute F1 test error, and that more than 50% of the examples need to be relabeled.
arXiv Detail & Related papers (2020-04-30T15:07:37Z) - iTAML: An Incremental Task-Agnostic Meta-learning Approach [123.10294801296926]
Humans can continuously learn new knowledge as their experience grows.
Previous learning in deep neural networks can quickly fade out when they are trained on a new task.
We introduce a novel meta-learning approach that seeks to maintain an equilibrium between all encountered tasks.
arXiv Detail & Related papers (2020-03-25T21:42:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.