CL-UZH at SemEval-2023 Task 10: Sexism Detection through Incremental
Fine-Tuning and Multi-Task Learning with Label Descriptions
- URL: http://arxiv.org/abs/2306.03907v1
- Date: Tue, 6 Jun 2023 17:59:49 GMT
- Title: CL-UZH at SemEval-2023 Task 10: Sexism Detection through Incremental
Fine-Tuning and Multi-Task Learning with Label Descriptions
- Authors: Janis Goldzycher
- Abstract summary: SemEval shared task textitTowards Explainable Detection of Online Sexism (EDOS 2023) is to detect sexism in English social media posts.
We present our submitted systems for all three subtasks, based on a multi-task model that has been fine-tuned on a range of related tasks.
We implement multi-task learning by formulating each task as binary pairwise text classification, where the dataset and label descriptions are given along with the input text.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The widespread popularity of social media has led to an increase in hateful,
abusive, and sexist language, motivating methods for the automatic detection of
such phenomena. The goal of the SemEval shared task \textit{Towards Explainable
Detection of Online Sexism} (EDOS 2023) is to detect sexism in English social
media posts (subtask A), and to categorize such posts into four coarse-grained
sexism categories (subtask B), and eleven fine-grained subcategories (subtask
C). In this paper, we present our submitted systems for all three subtasks,
based on a multi-task model that has been fine-tuned on a range of related
tasks and datasets before being fine-tuned on the specific EDOS subtasks. We
implement multi-task learning by formulating each task as binary pairwise text
classification, where the dataset and label descriptions are given along with
the input text. The results show clear improvements over a fine-tuned
DeBERTa-V3 serving as a baseline leading to $F_1$-scores of 85.9\% in subtask A
(rank 13/84), 64.8\% in subtask B (rank 19/69), and 44.9\% in subtask C
(26/63).
Related papers
- GenAI Content Detection Task 1: English and Multilingual Machine-Generated Text Detection: AI vs. Human [71.42669028683741]
We present a shared task on binary machine generated text detection conducted as a part of the GenAI workshop at COLING 2025.
The task consists of two subtasks: Monolingual (English) and Multilingual.
We provide a comprehensive overview of the data, a summary of the results, detailed descriptions of the participating systems, and an in-depth analysis of submissions.
arXiv Detail & Related papers (2025-01-19T11:11:55Z) - SemEval-2024 Task 8: Multidomain, Multimodel and Multilingual Machine-Generated Text Detection [68.858931667807]
Subtask A is a binary classification task determining whether a text is written by a human or generated by a machine.
Subtask B is to detect the exact source of a text, discerning whether it is written by a human or generated by a specific LLM.
Subtask C aims to identify the changing point within a text, at which the authorship transitions from human to machine.
arXiv Detail & Related papers (2024-04-22T13:56:07Z) - SemEval 2024 -- Task 10: Emotion Discovery and Reasoning its Flip in
Conversation (EDiReF) [61.49972925493912]
SemEval-2024 Task 10 is a shared task centred on identifying emotions in code-mixed dialogues.
This task comprises three distinct subtasks - emotion recognition in conversation for code-mixed dialogues, emotion flip reasoning for code-mixed dialogues, and emotion flip reasoning for English dialogues.
A total of 84 participants engaged in this task, with the most adept systems attaining F1-scores of 0.70, 0.79, and 0.76 for the respective subtasks.
arXiv Detail & Related papers (2024-02-29T08:20:06Z) - LCT-1 at SemEval-2023 Task 10: Pre-training and Multi-task Learning for
Sexism Detection and Classification [0.0]
SemEval-2023 Task 10 on Explainable Detection of Online Sexism aims at increasing explainability of the sexism detection.
Our system is based on further domain-adaptive pre-training.
In experiments, multi-task learning performs on par with standard fine-tuning for sexism detection.
arXiv Detail & Related papers (2023-06-08T09:56:57Z) - Multimodal Subtask Graph Generation from Instructional Videos [51.96856868195961]
Real-world tasks consist of multiple inter-dependent subtasks.
In this work, we aim to model the causal dependencies between such subtasks from instructional videos describing the task.
We present Multimodal Subtask Graph Generation (MSG2), an approach that constructs a Subtask Graph defining the dependency between a task's subtasks relevant to a task from noisy web videos.
arXiv Detail & Related papers (2023-02-17T03:41:38Z) - Overview of Abusive and Threatening Language Detection in Urdu at FIRE
2021 [50.591267188664666]
We present two shared tasks of abusive and threatening language detection for the Urdu language.
We present two manually annotated datasets containing tweets labelled as (i) Abusive and Non-Abusive, and (ii) Threatening and Non-Threatening.
For both subtasks, m-Bert based transformer model showed the best performance.
arXiv Detail & Related papers (2022-07-14T07:38:13Z) - Combating Hostility: Covid-19 Fake News and Hostile Post Detection in
Social Media [0.0]
This paper illustrates a detail description of the system and its results that developed as a part of the participation at CONSTRAINT shared task in AAAI-2021.
Various techniques are used to perform the classification task, including SVM, CNN, BiLSTM, and CNN+BiLSTM with tf-idf and Word2Vec embedding techniques.
arXiv Detail & Related papers (2021-01-09T05:15:41Z) - Ghmerti at SemEval-2019 Task 6: A Deep Word- and Character-based
Approach to Offensive Language Identification [1.192436948211501]
OffensEval addresses the problem of identifying and categorizing offensive language in social media.
The proposed approach includes character-level Convolutional Neural Network, word-level Recurrent Neural Network, and some preprocessing.
The performance achieved by the proposed model for subtask A is 77.93% macro-averaged F1-score.
arXiv Detail & Related papers (2020-09-22T20:13:48Z) - GUIR at SemEval-2020 Task 12: Domain-Tuned Contextualized Models for
Offensive Language Detection [27.45642971636561]
OffensEval 2020 task includes three English sub-tasks: identifying the presence of offensive language (Sub-task A), identifying the presence of target in offensive language (Sub-task B), and identifying the categories of the target (Sub-task C)
Our submissions achieve F1 scores of 91.7% in Sub-task A, 66.5% in Sub-task B, and 63.2% in Sub-task C.
arXiv Detail & Related papers (2020-07-28T20:45:43Z) - Kungfupanda at SemEval-2020 Task 12: BERT-Based Multi-Task Learning for
Offensive Language Detection [55.445023584632175]
We build an offensive language detection system, which combines multi-task learning with BERT-based models.
Our model achieves 91.51% F1 score in English Sub-task A, which is comparable to the first place.
arXiv Detail & Related papers (2020-04-28T11:27:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.