NutCracker at WNUT-2020 Task 2: Robustly Identifying Informative
COVID-19 Tweets using Ensembling and Adversarial Training
- URL: http://arxiv.org/abs/2010.04335v1
- Date: Fri, 9 Oct 2020 02:46:51 GMT
- Title: NutCracker at WNUT-2020 Task 2: Robustly Identifying Informative
COVID-19 Tweets using Ensembling and Adversarial Training
- Authors: Priyanshu Kumar and Aadarsh Singh
- Abstract summary: We experiment with COVID-Twitter-BERT and RoBERTa models to identify informative COVID-19 tweets.
The ensemble of COVID-Twitter-BERT and RoBERTa obtains a F1-score of 0.9096 on the test data of WNUT-2020 Task 2 and ranks 1st on the leaderboard.
- Score: 6.85316573653194
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We experiment with COVID-Twitter-BERT and RoBERTa models to identify
informative COVID-19 tweets. We further experiment with adversarial training to
make our models robust. The ensemble of COVID-Twitter-BERT and RoBERTa obtains
a F1-score of 0.9096 (on the positive class) on the test data of WNUT-2020 Task
2 and ranks 1st on the leaderboard. The ensemble of the models trained using
adversarial training also produces similar result.
Related papers
- Arena Learning: Build Data Flywheel for LLMs Post-training via Simulated Chatbot Arena [126.70522244144088]
We introduce Arena Learning, an innovative offline strategy designed to simulate arena battles using AI-driven annotations.
Arena Learning ensures precise evaluations and maintains consistency between offline simulations and online competitions.
We apply Arena Learning to train our target model, WizardLM-$beta$, and demonstrate significant performance enhancements.
arXiv Detail & Related papers (2024-07-15T11:26:07Z) - ThangDLU at #SMM4H 2024: Encoder-decoder models for classifying text data on social disorders in children and adolescents [49.00494558898933]
This paper describes our participation in Task 3 and Task 5 of the #SMM4H (Social Media Mining for Health) 2024 Workshop.
Task 3 is a multi-class classification task centered on tweets discussing the impact of outdoor environments on symptoms of social anxiety.
Task 5 involves a binary classification task focusing on tweets reporting medical disorders in children.
We applied transfer learning from pre-trained encoder-decoder models such as BART-base and T5-small to identify the labels of a set of given tweets.
arXiv Detail & Related papers (2024-04-30T17:06:20Z) - Outlier Robust Adversarial Training [57.06824365801612]
We introduce Outlier Robust Adversarial Training (ORAT) in this work.
ORAT is based on a bi-level optimization formulation of adversarial training with a robust rank-based loss function.
We show that the learning objective of ORAT satisfies the $mathcalH$-consistency in binary classification, which establishes it as a proper surrogate to adversarial 0/1 loss.
arXiv Detail & Related papers (2023-09-10T21:36:38Z) - Bag of Tricks for Effective Language Model Pretraining and Downstream
Adaptation: A Case Study on GLUE [93.98660272309974]
This report briefly describes our submission Vega v1 on the General Language Understanding Evaluation leaderboard.
GLUE is a collection of nine natural language understanding tasks, including question answering, linguistic acceptability, sentiment analysis, text similarity, paraphrase detection, and natural language inference.
With our optimized pretraining and fine-tuning strategies, our 1.3 billion model sets new state-of-the-art on 4/9 tasks, achieving the best average score of 91.3.
arXiv Detail & Related papers (2023-02-18T09:26:35Z) - Mutual Adversarial Training: Learning together is better than going
alone [82.78852509965547]
We study how interactions among models affect robustness via knowledge distillation.
We propose mutual adversarial training (MAT) in which multiple models are trained together.
MAT can effectively improve model robustness and outperform state-of-the-art methods under white-box attacks.
arXiv Detail & Related papers (2021-12-09T15:59:42Z) - Detection of COVID-19 informative tweets using RoBERTa [5.564705758320338]
We present our work to detect informative Covid-19 English tweets using RoBERTa model as a part of the W-NUT workshop 2020.
We show the efficacy of our model on a public dataset with an F1-score of 0.89 on the validation dataset and 0.87 on the leaderboard.
arXiv Detail & Related papers (2020-10-21T18:43:13Z) - CIA_NITT at WNUT-2020 Task 2: Classification of COVID-19 Tweets Using
Pre-trained Language Models [0.0]
We treat this as binary text classification problem and experiment with pre-trained language models.
Our first model which is based on CT-BERT achieves F1-score of 88.7% and second model which is ensemble of CT-BERT, RoBERTa and SVM achieves F1-score of 88.52%.
arXiv Detail & Related papers (2020-09-12T12:59:54Z) - LynyrdSkynyrd at WNUT-2020 Task 2: Semi-Supervised Learning for
Identification of Informative COVID-19 English Tweets [4.361526134899725]
We describe our system for WNUT-2020 shared task on the identification of informative COVID-19 English tweets.
Our system is an ensemble of various machine learning methods, leveraging both traditional feature-based classifiers as well as recent advances in pre-trained language models.
Our best performing model achieves an F1-score of 0.9179 on the provided validation set and 0.8805 on the blind test-set.
arXiv Detail & Related papers (2020-09-08T16:29:25Z) - UIT-HSE at WNUT-2020 Task 2: Exploiting CT-BERT for Identifying COVID-19
Information on the Twitter Social Network [2.7528170226206443]
In this paper, we present our results at the W-NUT 2020 Shared Task 2: Identification of Informative COVID-19 English Tweets.
We propose our simple but effective approach using the transformer-based models based on COVID-Twitter-BERT (CT-BERT) with different fine-tuning techniques.
As a result, we achieve the F1-Score of 90.94% with the third place on the leaderboard of this task which attracted 56 submitted teams in total.
arXiv Detail & Related papers (2020-09-07T08:20:31Z) - BANANA at WNUT-2020 Task 2: Identifying COVID-19 Information on Twitter
by Combining Deep Learning and Transfer Learning Models [0.0]
This paper describes our prediction system for WNUT-2020 Task 2: Identification of Informative COVID-19 English Tweets.
The dataset for this task contains size 10,000 tweets in English labeled by humans.
The experimental result indicates that we have achieved F1 for the INFORMATIVE label on our systems at 88.81% on the test set.
arXiv Detail & Related papers (2020-09-06T08:24:55Z) - Adversarial Robustness: From Self-Supervised Pre-Training to Fine-Tuning [134.15174177472807]
We introduce adversarial training into self-supervision, to provide general-purpose robust pre-trained models for the first time.
We conduct extensive experiments to demonstrate that the proposed framework achieves large performance margins.
arXiv Detail & Related papers (2020-03-28T18:28:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.