Team Enigma at ArgMining-EMNLP 2021: Leveraging Pre-trained Language
Models for Key Point Matching
- URL: http://arxiv.org/abs/2110.12370v1
- Date: Sun, 24 Oct 2021 07:10:39 GMT
- Title: Team Enigma at ArgMining-EMNLP 2021: Leveraging Pre-trained Language
Models for Key Point Matching
- Authors: Manav Nitin Kapadnis, Sohan Patnaik, Siba Smarak Panigrahi, Varun
Madhavan, Abhilash Nandy
- Abstract summary: We present the system description for our submission towards the Key Point Analysis Shared Task at ArgMining 2021.
We leveraged existing state of the art pre-trained language models along with incorporating additional data and features extracted from the inputs (topics, key points, and arguments) to improve performance.
We were able to achieve mAP strict and mAP relaxed score of 0.872 and 0.966 respectively in the evaluation phase, securing 5th place on the leaderboard.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present the system description for our submission towards the Key Point
Analysis Shared Task at ArgMining 2021. Track 1 of the shared task requires
participants to develop methods to predict the match score between each pair of
arguments and keypoints, provided they belong to the same topic under the same
stance. We leveraged existing state of the art pre-trained language models
along with incorporating additional data and features extracted from the inputs
(topics, key points, and arguments) to improve performance. We were able to
achieve mAP strict and mAP relaxed score of 0.872 and 0.966 respectively in the
evaluation phase, securing 5th place on the leaderboard. In the post evaluation
phase, we achieved a mAP strict and mAP relaxed score of 0.921 and 0.982
respectively. All the codes to generate reproducible results on our models are
available on Github.
Related papers
- Unify word-level and span-level tasks: NJUNLP's Participation for the
WMT2023 Quality Estimation Shared Task [59.46906545506715]
We introduce the NJUNLP team to the WMT 2023 Quality Estimation (QE) shared task.
Our team submitted predictions for the English-German language pair on all two sub-tasks.
Our models achieved the best results in English-German for both word-level and fine-grained error span detection sub-tasks.
arXiv Detail & Related papers (2023-09-23T01:52:14Z) - Selective In-Context Data Augmentation for Intent Detection using
Pointwise V-Information [100.03188187735624]
We introduce a novel approach based on PLMs and pointwise V-information (PVI), a metric that can measure the usefulness of a datapoint for training a model.
Our method first fine-tunes a PLM on a small seed of training data and then synthesizes new datapoints - utterances that correspond to given intents.
Our method is thus able to leverage the expressive power of large language models to produce diverse training data.
arXiv Detail & Related papers (2023-02-10T07:37:49Z) - Unifying Language Learning Paradigms [96.35981503087567]
We present a unified framework for pre-training models that are universally effective across datasets and setups.
We show how different pre-training objectives can be cast as one another and how interpolating between different objectives can be effective.
Our model also achieve strong results at in-context learning, outperforming 175B GPT-3 on zero-shot SuperGLUE and tripling the performance of T5-XXL on one-shot summarization.
arXiv Detail & Related papers (2022-05-10T19:32:20Z) - End-to-End Zero-Shot HOI Detection via Vision and Language Knowledge
Distillation [86.41437210485932]
We aim at advancing zero-shot HOI detection to detect both seen and unseen HOIs simultaneously.
We propose a novel end-to-end zero-shot HOI Detection framework via vision-language knowledge distillation.
Our method outperforms the previous SOTA by 8.92% on unseen mAP and 10.18% on overall mAP.
arXiv Detail & Related papers (2022-04-01T07:27:19Z) - NLP-CIC @ PRELEARN: Mastering prerequisites relations, from handcrafted
features to embeddings [68.97335984455059]
We present our systems and findings for the prerequisite relation learning task (PRELEARN) at EVALITA 2020.
The task aims to classify whether a pair of concepts hold a prerequisite relation or not.
Our submissions ranked first place in both scenarios with average F1 score of 0.887 and 0.690 respectively across domains on the test sets.
arXiv Detail & Related papers (2020-11-07T12:13:09Z) - PublishInCovid19 at WNUT 2020 Shared Task-1: Entity Recognition in Wet
Lab Protocols using Structured Learning Ensemble and Contextualised
Embeddings [0.0]
We describe the approach that we employed to address the task of Entity Recognition over Wet Lab Protocols.
In the first phase, we experiment with various contextualised word embeddings and a BiLSTM-CRF model.
In the second phase, we create an ensemble composed of eleven BiLSTM-CRF models.
Our final submission achieved a micro F1-score of 0.8175 and 0.7757 for the partial and exact match of the entity spans.
arXiv Detail & Related papers (2020-10-05T16:45:30Z) - Phonemer at WNUT-2020 Task 2: Sequence Classification Using COVID
Twitter BERT and Bagging Ensemble Technique based on Plurality Voting [0.0]
We develop a system that automatically identifies whether an English Tweet related to the novel coronavirus (COVID-19) is informative or not.
Our final approach achieved an F1-score of 0.9037 and we were ranked sixth overall with F1-score as the evaluation criteria.
arXiv Detail & Related papers (2020-10-01T10:54:54Z) - FiSSA at SemEval-2020 Task 9: Fine-tuned For Feelings [2.362412515574206]
In this paper, we present our approach for sentiment classification on Spanish-English code-mixed social media data.
We explore both monolingual and multilingual models with the standard fine-tuning method.
Although two-step fine-tuning improves sentiment classification performance over the base model, the large multilingual XLM-RoBERTa model achieves best weighted F1-score.
arXiv Detail & Related papers (2020-07-24T14:48:27Z) - Voice@SRIB at SemEval-2020 Task 9 and 12: Stacked Ensembling method for
Sentiment and Offensiveness detection in Social Media [2.9008108937701333]
We train embeddings, ensembling methods for Sentimix, and OffensEval tasks.
We evaluate our models on macro F1-score, precision, accuracy, and recall on the datasets.
arXiv Detail & Related papers (2020-07-20T11:54:43Z) - AMR Parsing via Graph-Sequence Iterative Inference [62.85003739964878]
We propose a new end-to-end model that treats AMR parsing as a series of dual decisions on the input sequence and the incrementally constructed graph.
We show that the answers to these two questions are mutually causalities.
We design a model based on iterative inference that helps achieve better answers in both perspectives, leading to greatly improved parsing accuracy.
arXiv Detail & Related papers (2020-04-12T09:15:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.