Hopeful_Men@LT-EDI-EACL2021: Hope Speech Detection Using Indic
Transliteration and Transformers
- URL: http://arxiv.org/abs/2102.12082v2
- Date: Thu, 25 Feb 2021 04:50:47 GMT
- Title: Hopeful_Men@LT-EDI-EACL2021: Hope Speech Detection Using Indic
Transliteration and Transformers
- Authors: Ishan Sanjeev Upadhyay, Nikhil E, Anshul Wadhawan, Radhika Mamidi
- Abstract summary: This paper describes the approach we used to detect hope speech in the HopeEDI dataset.
In the first approach, we used contextual embeddings to train classifiers using logistic regression, random forest, SVM, and LSTM based models.
The second approach involved using a majority voting ensemble of 11 models which were obtained by fine-tuning pre-trained transformer models.
- Score: 6.955778726800376
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper aims to describe the approach we used to detect hope speech in the
HopeEDI dataset. We experimented with two approaches. In the first approach, we
used contextual embeddings to train classifiers using logistic regression,
random forest, SVM, and LSTM based models.The second approach involved using a
majority voting ensemble of 11 models which were obtained by fine-tuning
pre-trained transformer models (BERT, ALBERT, RoBERTa, IndicBERT) after adding
an output layer. We found that the second approach was superior for English,
Tamil and Malayalam. Our solution got a weighted F1 score of 0.93, 0.75 and
0.49 for English,Malayalam and Tamil respectively. Our solution ranked first in
English, eighth in Malayalam and eleventh in Tamil.
Related papers
- CTC-based Non-autoregressive Speech Translation [51.37920141751813]
We investigate the potential of connectionist temporal classification for non-autoregressive speech translation.
We develop a model consisting of two encoders that are guided by CTC to predict the source and target texts.
Experiments on the MuST-C benchmarks show that our NAST model achieves an average BLEU score of 29.5 with a speed-up of 5.67$times$.
arXiv Detail & Related papers (2023-05-27T03:54:09Z) - Bag of Tricks for Effective Language Model Pretraining and Downstream
Adaptation: A Case Study on GLUE [93.98660272309974]
This report briefly describes our submission Vega v1 on the General Language Understanding Evaluation leaderboard.
GLUE is a collection of nine natural language understanding tasks, including question answering, linguistic acceptability, sentiment analysis, text similarity, paraphrase detection, and natural language inference.
With our optimized pretraining and fine-tuning strategies, our 1.3 billion model sets new state-of-the-art on 4/9 tasks, achieving the best average score of 91.3.
arXiv Detail & Related papers (2023-02-18T09:26:35Z) - Ensemble Transfer Learning for Multilingual Coreference Resolution [60.409789753164944]
A problem that frequently occurs when working with a non-English language is the scarcity of annotated training data.
We design a simple but effective ensemble-based framework that combines various transfer learning techniques.
We also propose a low-cost TL method that bootstraps coreference resolution models by utilizing Wikipedia anchor texts.
arXiv Detail & Related papers (2023-01-22T18:22:55Z) - M3ST: Mix at Three Levels for Speech Translation [66.71994367650461]
We propose Mix at three levels for Speech Translation (M3ST) method to increase the diversity of the augmented training corpus.
In the first stage of fine-tuning, we mix the training corpus at three levels, including word level, sentence level and frame level, and fine-tune the entire model with mixed data.
Experiments on MuST-C speech translation benchmark and analysis show that M3ST outperforms current strong baselines and achieves state-of-the-art results on eight directions with an average BLEU of 29.9.
arXiv Detail & Related papers (2022-12-07T14:22:00Z) - BJTU-WeChat's Systems for the WMT22 Chat Translation Task [66.81525961469494]
This paper introduces the joint submission of the Beijing Jiaotong University and WeChat AI to the WMT'22 chat translation task for English-German.
Based on the Transformer, we apply several effective variants.
Our systems achieve 0.810 and 0.946 COMET scores.
arXiv Detail & Related papers (2022-11-28T02:35:04Z) - Towards Arabic Sentence Simplification via Classification and Generative
Approaches [0.0]
This paper presents an attempt to build a Modern Standard Arabic (MSA) sentence-level simplification system.
We experimented with sentence simplification using two approaches: (i) a classification approach leading to lexical simplification pipelines which use Arabic-BERT, a pre-trained contextualised model, as well as a model of fastText word embeddings; and (ii) a generative approach, a Seq2Seq technique by applying a multilingual Text-to-Text Transfer Transformer mT5.
arXiv Detail & Related papers (2022-04-20T08:17:33Z) - PSG@HASOC-Dravidian CodeMixFIRE2021: Pretrained Transformers for
Offensive Language Identification in Tanglish [0.0]
This paper describes the system submitted to Dravidian-Codemix-HASOC2021: Hate Speech and Offensive Language Identification in Dravidian languages.
This task aims to identify offensive content in code-mixed comments/posts in Dravidian languages collected from social media.
arXiv Detail & Related papers (2021-10-06T15:23:40Z) - Cross-Lingual Text Classification of Transliterated Hindi and Malayalam [31.86825573676501]
We combine data augmentation approaches with a Teacher-Student training scheme to address this issue.
We evaluate our method on transliterated Hindi and Malayalam, also introducing new datasets for benchmarking on real-world scenarios.
Our method yielded an average improvement of +5.6% on mBERT and +4.7% on XLM-R in F1 scores over their strong baselines.
arXiv Detail & Related papers (2021-08-31T05:13:17Z) - The USYD-JD Speech Translation System for IWSLT 2021 [85.64797317290349]
This paper describes the University of Sydney& JD's joint submission of the IWSLT 2021 low resource speech translation task.
We trained our models with the officially provided ASR and MT datasets.
To achieve better translation performance, we explored the most recent effective strategies, including back translation, knowledge distillation, multi-feature reranking and transductive finetuning.
arXiv Detail & Related papers (2021-07-24T09:53:34Z) - LaDiff ULMFiT: A Layer Differentiated training approach for ULMFiT [0.0]
We propose a Layer Differentiated training procedure for training a pre-trained ULMFiT arXiv:1801.06146 model.
We used special tokens to annotate specific parts of the tweets to improve language understanding and gain insights on the model.
The proposed approach ranked 61st out of 164 in the sub-task "COVID19 Fake News Detection in English"
arXiv Detail & Related papers (2021-01-13T09:52:04Z) - HinglishNLP: Fine-tuned Language Models for Hinglish Sentiment Detection [0.0]
This work adds two common approaches to sentiment analysis.
NB-SVM beats RoBERTa by 6.2% (relative) F1.
The best performing model is a majority-vote ensemble which achieves an F1 of 0.707.
arXiv Detail & Related papers (2020-08-22T12:01:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.