CIA_NITT at WNUT-2020 Task 2: Classification of COVID-19 Tweets Using
Pre-trained Language Models
- URL: http://arxiv.org/abs/2009.05782v1
- Date: Sat, 12 Sep 2020 12:59:54 GMT
- Title: CIA_NITT at WNUT-2020 Task 2: Classification of COVID-19 Tweets Using
Pre-trained Language Models
- Authors: Yandrapati Prakash Babu and Rajagopal Eswari
- Abstract summary: We treat this as binary text classification problem and experiment with pre-trained language models.
Our first model which is based on CT-BERT achieves F1-score of 88.7% and second model which is ensemble of CT-BERT, RoBERTa and SVM achieves F1-score of 88.52%.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper presents our models for WNUT 2020 shared task2. The shared task2
involves identification of COVID-19 related informative tweets. We treat this
as binary text classification problem and experiment with pre-trained language
models. Our first model which is based on CT-BERT achieves F1-score of 88.7%
and second model which is an ensemble of CT-BERT, RoBERTa and SVM achieves
F1-score of 88.52%.
Related papers
- Text Augmentations with R-drop for Classification of Tweets Self
Reporting Covid-19 [28.91836510067532]
This paper presents models created for the Social Media Mining for Health 2023 shared task.
Our approach involves a classification model that incorporates diverse textual augmentations.
Our system achieves an impressive F1 score of 0.877 on the test set.
arXiv Detail & Related papers (2023-11-06T14:18:16Z) - BJTU-WeChat's Systems for the WMT22 Chat Translation Task [66.81525961469494]
This paper introduces the joint submission of the Beijing Jiaotong University and WeChat AI to the WMT'22 chat translation task for English-German.
Based on the Transformer, we apply several effective variants.
Our systems achieve 0.810 and 0.946 COMET scores.
arXiv Detail & Related papers (2022-11-28T02:35:04Z) - RoBLEURT Submission for the WMT2021 Metrics Task [72.26898579202076]
We present our submission to the Shared Metrics Task: RoBLEURT.
Our model reaches state-of-the-art correlations with the WMT 2020 human annotations upon 8 out of 10 to-English language pairs.
arXiv Detail & Related papers (2022-04-28T08:49:40Z) - Is BERT a Cross-Disciplinary Knowledge Learner? A Surprising Finding of
Pre-trained Models' Transferability [74.11825654535895]
We investigate whether the power of the models pre-trained on text data, such as BERT, can be transferred to general token sequence classification applications.
We find that even on non-text data, the models pre-trained on text converge faster than the randomly models.
arXiv Detail & Related papers (2021-03-12T09:19:14Z) - KgPLM: Knowledge-guided Language Model Pre-training via Generative and
Discriminative Learning [45.067001062192844]
We present a language model pre-training framework guided by factual knowledge completion and verification.
Experimental results on LAMA, a set of zero-shot cloze-style question answering tasks, show that our model contains richer factual knowledge than the conventional pre-trained language models.
arXiv Detail & Related papers (2020-12-07T09:39:25Z) - NIT COVID-19 at WNUT-2020 Task 2: Deep Learning Model RoBERTa for
Identify Informative COVID-19 English Tweets [0.0]
This paper presents the model submitted by the NIT_COVID-19 team for identified informative COVID-19 English tweets at WNUT-2020 Task2.
The performance achieved by the proposed model for shared task WNUT 2020 Task2 is 89.14% in the F1-score metric.
arXiv Detail & Related papers (2020-11-11T05:20:39Z) - NLP-CIC @ PRELEARN: Mastering prerequisites relations, from handcrafted
features to embeddings [68.97335984455059]
We present our systems and findings for the prerequisite relation learning task (PRELEARN) at EVALITA 2020.
The task aims to classify whether a pair of concepts hold a prerequisite relation or not.
Our submissions ranked first place in both scenarios with average F1 score of 0.887 and 0.690 respectively across domains on the test sets.
arXiv Detail & Related papers (2020-11-07T12:13:09Z) - NutCracker at WNUT-2020 Task 2: Robustly Identifying Informative
COVID-19 Tweets using Ensembling and Adversarial Training [6.85316573653194]
We experiment with COVID-Twitter-BERT and RoBERTa models to identify informative COVID-19 tweets.
The ensemble of COVID-Twitter-BERT and RoBERTa obtains a F1-score of 0.9096 on the test data of WNUT-2020 Task 2 and ranks 1st on the leaderboard.
arXiv Detail & Related papers (2020-10-09T02:46:51Z) - Phonemer at WNUT-2020 Task 2: Sequence Classification Using COVID
Twitter BERT and Bagging Ensemble Technique based on Plurality Voting [0.0]
We develop a system that automatically identifies whether an English Tweet related to the novel coronavirus (COVID-19) is informative or not.
Our final approach achieved an F1-score of 0.9037 and we were ranked sixth overall with F1-score as the evaluation criteria.
arXiv Detail & Related papers (2020-10-01T10:54:54Z) - LynyrdSkynyrd at WNUT-2020 Task 2: Semi-Supervised Learning for
Identification of Informative COVID-19 English Tweets [4.361526134899725]
We describe our system for WNUT-2020 shared task on the identification of informative COVID-19 English tweets.
Our system is an ensemble of various machine learning methods, leveraging both traditional feature-based classifiers as well as recent advances in pre-trained language models.
Our best performing model achieves an F1-score of 0.9179 on the provided validation set and 0.8805 on the blind test-set.
arXiv Detail & Related papers (2020-09-08T16:29:25Z) - DeBERTa: Decoding-enhanced BERT with Disentangled Attention [119.77305080520718]
We propose a new model architecture DeBERTa that improves the BERT and RoBERTa models using two novel techniques.
We show that these techniques significantly improve the efficiency of model pre-training and the performance of both natural language understanding (NLU) and natural langauge generation (NLG) downstream tasks.
arXiv Detail & Related papers (2020-06-05T19:54:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.