Plumeria at SemEval-2022 Task 6: Robust Approaches for Sarcasm Detection
for English and Arabic Using Transformers and Data Augmentation
- URL: http://arxiv.org/abs/2203.04111v1
- Date: Tue, 8 Mar 2022 14:33:45 GMT
- Title: Plumeria at SemEval-2022 Task 6: Robust Approaches for Sarcasm Detection
for English and Arabic Using Transformers and Data Augmentation
- Authors: Shubham Kumar Nigam and Mosab Shaheen
- Abstract summary: This paper describes our submission to SemEval-2022 Task 6 on sarcasm detection and its five subtasks for English and Arabic.
For detecting sarcasm, we used deep learning techniques based on transformers due to its success in the field of Natural Language Processing (NLP) without the need for feature engineering.
- Score: 0.951828574518325
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper describes our submission to SemEval-2022 Task 6 on sarcasm
detection and its five subtasks for English and Arabic. Sarcasm conveys a
meaning which contradicts the literal meaning, and it is mainly found on social
networks. It has a significant role in understanding the intention of the user.
For detecting sarcasm, we used deep learning techniques based on transformers
due to its success in the field of Natural Language Processing (NLP) without
the need for feature engineering. The datasets were taken from tweets. We
created new datasets by augmenting with external data or by using word
embeddings and repetition of instances. Experiments were done on the datasets
with different types of preprocessing because it is crucial in this task. The
rank of our team was consistent across four subtasks (fourth rank in three
subtasks and sixth rank in one subtask); whereas other teams might be in the
top ranks for some subtasks but rank drastically less in other subtasks. This
implies the robustness and stability of the models and the techniques we used.
Related papers
- Rethinking and Improving Multi-task Learning for End-to-end Speech
Translation [51.713683037303035]
We investigate the consistency between different tasks, considering different times and modules.
We find that the textual encoder primarily facilitates cross-modal conversion, but the presence of noise in speech impedes the consistency between text and speech representations.
We propose an improved multi-task learning (IMTL) approach for the ST task, which bridges the modal gap by mitigating the difference in length and representation.
arXiv Detail & Related papers (2023-11-07T08:48:46Z) - Findings of the WMT 2022 Shared Task on Translation Suggestion [63.457874930232926]
We report the result of the first edition of the WMT shared task on Translation Suggestion.
The task aims to provide alternatives for specific words or phrases given the entire documents generated by machine translation (MT)
It consists two sub-tasks, namely, the naive translation suggestion and translation suggestion with hints.
arXiv Detail & Related papers (2022-11-30T03:48:36Z) - Sarcasm Detection Framework Using Emotion and Sentiment Features [62.997667081978825]
We propose a model which incorporates emotion and sentiment features to capture the incongruity intrinsic to sarcasm.
Our approach achieved state-of-the-art results on four datasets from social networking platforms and online media.
arXiv Detail & Related papers (2022-11-23T15:14:44Z) - Effective Cross-Task Transfer Learning for Explainable Natural Language
Inference with T5 [50.574918785575655]
We compare sequential fine-tuning with a model for multi-task learning in the context of boosting performance on two tasks.
Our results show that while sequential multi-task learning can be tuned to be good at the first of two target tasks, it performs less well on the second and additionally struggles with overfitting.
arXiv Detail & Related papers (2022-10-31T13:26:08Z) - CS-UM6P at SemEval-2022 Task 6: Transformer-based Models for Intended
Sarcasm Detection in English and Arabic [6.221019624345408]
Sarcasm is a form of figurative language where the intended meaning of a sentence differs from its literal meaning.
In this paper, we present our participating system to the intended sarcasm detection task in English and Arabic languages.
arXiv Detail & Related papers (2022-06-16T19:14:54Z) - Learning Language-Conditioned Robot Behavior from Offline Data and
Crowd-Sourced Annotation [80.29069988090912]
We study the problem of learning a range of vision-based manipulation tasks from a large offline dataset of robot interaction.
We propose to leverage offline robot datasets with crowd-sourced natural language labels.
We find that our approach outperforms both goal-image specifications and language conditioned imitation techniques by more than 25%.
arXiv Detail & Related papers (2021-09-02T17:42:13Z) - Sarcasm Detection in Twitter -- Performance Impact when using Data
Augmentation: Word Embeddings [0.0]
Sarcasm is the use of words usually used to either mock or annoy someone, or for humorous purposes.
We propose a contextual model for sarcasm identification in twitter using RoBERTa and augmenting the dataset.
We achieve performance gain by 3.2% in the iSarcasm dataset when using data augmentation to increase 20% of data labeled as sarcastic.
arXiv Detail & Related papers (2021-08-23T04:24:12Z) - MagicPai at SemEval-2021 Task 7: Method for Detecting and Rating Humor
Based on Multi-Task Adversarial Training [4.691435917434472]
This paper describes MagicPai's system for SemEval 2021 Task 7, HaHackathon: Detecting and Rating Humor and Offense.
This task aims to detect whether the text is humorous and how humorous it is.
We mainly present our solution, a multi-task learning model based on adversarial examples.
arXiv Detail & Related papers (2021-04-21T03:23:02Z) - Deep Learning for Text Style Transfer: A Survey [71.8870854396927]
Text style transfer is an important task in natural language generation, which aims to control certain attributes in the generated text.
We present a systematic survey of the research on neural text style transfer, spanning over 100 representative articles since the first neural text style transfer work in 2017.
We discuss the task formulation, existing datasets and subtasks, evaluation, as well as the rich methodologies in the presence of parallel and non-parallel data.
arXiv Detail & Related papers (2020-11-01T04:04:43Z) - IITK-RSA at SemEval-2020 Task 5: Detecting Counterfactuals [3.0396370700420063]
This paper describes our efforts in tackling Task 5 of SemEval-2020.
The task involved detecting a class of textual expressions known as counterfactuals.
Counterfactual statements describe events that have not or could not have occurred and the possible implications of such events.
arXiv Detail & Related papers (2020-07-21T14:45:53Z) - Sarcasm Detection using Context Separators in Online Discourse [3.655021726150369]
Sarcasm is an intricate form of speech, where meaning is conveyed implicitly.
In this work, we use RoBERTa_large to detect sarcasm in two datasets.
We also assert the importance of context in improving the performance of contextual word embedding models.
arXiv Detail & Related papers (2020-06-01T10:52:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.