tmn at SemEval-2023 Task 9: Multilingual Tweet Intimacy Detection using
XLM-T, Google Translate, and Ensemble Learning
- URL: http://arxiv.org/abs/2304.04054v1
- Date: Sat, 8 Apr 2023 15:50:16 GMT
- Title: tmn at SemEval-2023 Task 9: Multilingual Tweet Intimacy Detection using
XLM-T, Google Translate, and Ensemble Learning
- Authors: Anna Glazkova
- Abstract summary: The paper describes a transformer-based system designed for SemEval-2023 Task 9: Multilingual Tweet Intimacy Analysis.
The purpose of the task was to predict the intimacy of tweets in a range from 1 (not intimate at all) to 5 (very intimate)
- Score: 2.28438857884398
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The paper describes a transformer-based system designed for SemEval-2023 Task
9: Multilingual Tweet Intimacy Analysis. The purpose of the task was to predict
the intimacy of tweets in a range from 1 (not intimate at all) to 5 (very
intimate). The official training set for the competition consisted of tweets in
six languages (English, Spanish, Italian, Portuguese, French, and Chinese). The
test set included the given six languages as well as external data with four
languages not presented in the training set (Hindi, Arabic, Dutch, and Korean).
We presented a solution based on an ensemble of XLM-T, a multilingual RoBERTa
model adapted to the Twitter domain. To improve the performance of unseen
languages, each tweet was supplemented by its English translation. We explored
the effectiveness of translated data for the languages seen in fine-tuning
compared to unseen languages and estimated strategies for using translated data
in transformer-based models. Our solution ranked 4th on the leaderboard while
achieving an overall Pearson's r of 0.599 over the test set. The proposed
system improves up to 0.088 Pearson's r over a score averaged across all 45
submissions.
Related papers
- Decoupled Vocabulary Learning Enables Zero-Shot Translation from Unseen Languages [55.157295899188476]
neural machine translation systems learn to map sentences of different languages into a common representation space.
In this work, we test this hypothesis by zero-shot translating from unseen languages.
We demonstrate that this setup enables zero-shot translation from entirely unseen languages.
arXiv Detail & Related papers (2024-08-05T07:58:58Z) - Breaking Language Barriers in Multilingual Mathematical Reasoning: Insights and Observations [59.056367787688146]
This paper pioneers exploring and training powerful Multilingual Math Reasoning (xMR) LLMs.
We construct the first multilingual math reasoning instruction dataset, MGSM8KInstruct, encompassing ten distinct languages.
By utilizing translation, we construct the first multilingual math reasoning instruction dataset, MGSM8KInstruct, encompassing ten distinct languages.
arXiv Detail & Related papers (2023-10-31T08:09:20Z) - OPI at SemEval 2023 Task 9: A Simple But Effective Approach to
Multilingual Tweet Intimacy Analysis [0.0]
This paper describes our submission to the SemEval 2023 multilingual tweet intimacy analysis shared task.
The goal of the task was to assess the level of intimacy of Twitter posts in ten languages.
Our method was ranked first in five out of ten language subtasks, obtaining the highest average score across all languages.
arXiv Detail & Related papers (2023-04-14T13:49:28Z) - HULAT at SemEval-2023 Task 9: Data augmentation for pre-trained
transformers applied to Multilingual Tweet Intimacy Analysis [1.4213973379473652]
This paper describes our participation in SemEval-2023 Task 9, Intimacy Analysis of Multilingual Tweets.
We fine-tune some of the most popular transformer models with the training dataset and synthetic data generated by different data augmentation techniques.
Despite its modest results, our system shows promising results in languages such as Portuguese, English, and Dutch.
arXiv Detail & Related papers (2023-02-24T18:10:37Z) - Few-shot Learning with Multilingual Language Models [66.49496434282564]
We train multilingual autoregressive language models on a balanced corpus covering a diverse set of languages.
Our largest model sets new state of the art in few-shot learning in more than 20 representative languages.
We present a detailed analysis of where the model succeeds and fails, showing in particular that it enables cross-lingual in-context learning.
arXiv Detail & Related papers (2021-12-20T16:52:35Z) - Improved Multilingual Language Model Pretraining for Social Media Text
via Translation Pair Prediction [1.14219428942199]
We evaluate a simple approach to improving zero-shot multilingual transfer of mBERT on social media corpus.
Our approach assumes access to translations between source-target language pairs.
We show improvements from TPP pretraining over mBERT alone in zero-shot transfer from English to Hindi, Arabic, and Japanese.
arXiv Detail & Related papers (2021-10-20T00:06:26Z) - AmericasNLI: Evaluating Zero-shot Natural Language Understanding of
Pretrained Multilingual Models in Truly Low-resource Languages [75.08199398141744]
We present AmericasNLI, an extension of XNLI (Conneau et al.), to 10 indigenous languages of the Americas.
We conduct experiments with XLM-R, testing multiple zero-shot and translation-based approaches.
We find that XLM-R's zero-shot performance is poor for all 10 languages, with an average performance of 38.62%.
arXiv Detail & Related papers (2021-04-18T05:32:28Z) - Facebook AI's WMT20 News Translation Task Submission [69.92594751788403]
This paper describes Facebook AI's submission to WMT20 shared news translation task.
We focus on the low resource setting and participate in two language pairs, Tamil -> English and Inuktitut -> English.
We approach the low resource problem using two main strategies, leveraging all available data and adapting the system to the target news domain.
arXiv Detail & Related papers (2020-11-16T21:49:00Z) - UPB at SemEval-2020 Task 12: Multilingual Offensive Language Detection
on Social Media by Fine-tuning a Variety of BERT-based Models [0.0]
This paper describes our Transformer-based solutions for identifying offensive language on Twitter in five languages.
It was employed in Subtask A of the Offenseval 2020 shared task.
arXiv Detail & Related papers (2020-10-26T14:28:29Z) - Beyond English-Centric Multilingual Machine Translation [74.21727842163068]
We create a true Many-to-Many multilingual translation model that can translate directly between any pair of 100 languages.
We build and open source a training dataset that covers thousands of language directions with supervised data, created through large-scale mining.
Our focus on non-English-Centric models brings gains of more than 10 BLEU when directly translating between non-English directions while performing competitively to the best single systems of WMT.
arXiv Detail & Related papers (2020-10-21T17:01:23Z) - Cross-lingual Inductive Transfer to Detect Offensive Language [3.655021726150369]
We introduce a cross-lingual inductive approach to identify the offensive language in tweets using the contextual word embedding textitXLM-RoBERTa (XLM-R)
We show that our model performs competitively on all five languages.
arXiv Detail & Related papers (2020-07-07T20:10:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.