PublishInCovid19 at WNUT 2020 Shared Task-1: Entity Recognition in Wet
Lab Protocols using Structured Learning Ensemble and Contextualised
Embeddings
- URL: http://arxiv.org/abs/2010.02142v2
- Date: Thu, 15 Oct 2020 08:33:45 GMT
- Title: PublishInCovid19 at WNUT 2020 Shared Task-1: Entity Recognition in Wet
Lab Protocols using Structured Learning Ensemble and Contextualised
Embeddings
- Authors: Janvijay Singh, Anshul Wadhawan
- Abstract summary: We describe the approach that we employed to address the task of Entity Recognition over Wet Lab Protocols.
In the first phase, we experiment with various contextualised word embeddings and a BiLSTM-CRF model.
In the second phase, we create an ensemble composed of eleven BiLSTM-CRF models.
Our final submission achieved a micro F1-score of 0.8175 and 0.7757 for the partial and exact match of the entity spans.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we describe the approach that we employed to address the task
of Entity Recognition over Wet Lab Protocols -- a shared task in EMNLP
WNUT-2020 Workshop. Our approach is composed of two phases. In the first phase,
we experiment with various contextualised word embeddings (like Flair,
BERT-based) and a BiLSTM-CRF model to arrive at the best-performing
architecture. In the second phase, we create an ensemble composed of eleven
BiLSTM-CRF models. The individual models are trained on random train-validation
splits of the complete dataset. Here, we also experiment with different output
merging schemes, including Majority Voting and Structured Learning Ensembling
(SLE). Our final submission achieved a micro F1-score of 0.8175 and 0.7757 for
the partial and exact match of the entity spans, respectively. We were ranked
first and second, in terms of partial and exact match, respectively.
Related papers
- SAM as the Guide: Mastering Pseudo-Label Refinement in Semi-Supervised Referring Expression Segmentation [66.92696817276288]
SemiRES is a semi-supervised framework that effectively leverages a combination of labeled and unlabeled data to perform RES.
SemiRES incorporates the Segment Anything Model (SAM), renowned for its precise boundary demarcation.
In instances where a precise mask cannot be matched from the available candidates, we develop the Pixel-Wise Adjustment (PWA) strategy.
arXiv Detail & Related papers (2024-06-03T15:42:30Z) - Large Language Models aren't all that you need [0.0]
This paper describes the architecture and systems built towards solving the SemEval 2023 Task 2: MultiCoNER II.
We evaluate two approaches (a) a traditional Random Fields model and (b) a Large Language Model (LLM) fine-tuned with a customized head and compare the two approaches.
arXiv Detail & Related papers (2024-01-01T08:32:50Z) - Alibaba-Translate China's Submission for WMT 2022 Metrics Shared Task [61.34108034582074]
We build our system based on the core idea of UNITE (Unified Translation Evaluation)
During the model pre-training phase, we first apply the pseudo-labeled data examples to continuously pre-train UNITE.
During the fine-tuning phase, we use both Direct Assessment (DA) and Multidimensional Quality Metrics (MQM) data from past years' WMT competitions.
arXiv Detail & Related papers (2022-10-18T08:51:25Z) - Learning with MISELBO: The Mixture Cookbook [62.75516608080322]
We present the first ever mixture of variational approximations for a normalizing flow-based hierarchical variational autoencoder (VAE) with VampPrior and a PixelCNN decoder network.
We explain this cooperative behavior by drawing a novel connection between VI and adaptive importance sampling.
We obtain state-of-the-art results among VAE architectures in terms of negative log-likelihood on the MNIST and FashionMNIST datasets.
arXiv Detail & Related papers (2022-09-30T15:01:35Z) - COLO: A Contrastive Learning based Re-ranking Framework for One-Stage
Summarization [84.70895015194188]
We propose a Contrastive Learning based re-ranking framework for one-stage summarization called COLO.
COLO boosts the extractive and abstractive results of one-stage systems on CNN/DailyMail benchmark to 44.58 and 46.33 ROUGE-1 score.
arXiv Detail & Related papers (2022-09-29T06:11:21Z) - Federated Hetero-Task Learning [42.985155807178685]
We present B-FHTL, a federated hetero-task learning benchmark consisted of simulation dataset, FL protocols and a unified evaluation mechanism.
To ensure fair comparison among different FL algorithms, B-FHTL builds in a full suite of FL protocols.
We compare the FL algorithms in fields of federated multi-task learning, federated personalization and federated meta learning within B-FHTL.
arXiv Detail & Related papers (2022-06-07T16:43:09Z) - Team Enigma at ArgMining-EMNLP 2021: Leveraging Pre-trained Language
Models for Key Point Matching [0.0]
We present the system description for our submission towards the Key Point Analysis Shared Task at ArgMining 2021.
We leveraged existing state of the art pre-trained language models along with incorporating additional data and features extracted from the inputs (topics, key points, and arguments) to improve performance.
We were able to achieve mAP strict and mAP relaxed score of 0.872 and 0.966 respectively in the evaluation phase, securing 5th place on the leaderboard.
arXiv Detail & Related papers (2021-10-24T07:10:39Z) - Partner-Assisted Learning for Few-Shot Image Classification [54.66864961784989]
Few-shot Learning has been studied to mimic human visual capabilities and learn effective models without the need of exhaustive human annotation.
In this paper, we focus on the design of training strategy to obtain an elemental representation such that the prototype of each novel class can be estimated from a few labeled samples.
We propose a two-stage training scheme, which first trains a partner encoder to model pair-wise similarities and extract features serving as soft-anchors, and then trains a main encoder by aligning its outputs with soft-anchors while attempting to maximize classification performance.
arXiv Detail & Related papers (2021-09-15T22:46:19Z) - Few-Shot Named Entity Recognition: A Comprehensive Study [92.40991050806544]
We investigate three schemes to improve the model generalization ability for few-shot settings.
We perform empirical comparisons on 10 public NER datasets with various proportions of labeled data.
We create new state-of-the-art results on both few-shot and training-free settings.
arXiv Detail & Related papers (2020-12-29T23:43:16Z) - Phonemer at WNUT-2020 Task 2: Sequence Classification Using COVID
Twitter BERT and Bagging Ensemble Technique based on Plurality Voting [0.0]
We develop a system that automatically identifies whether an English Tweet related to the novel coronavirus (COVID-19) is informative or not.
Our final approach achieved an F1-score of 0.9037 and we were ranked sixth overall with F1-score as the evaluation criteria.
arXiv Detail & Related papers (2020-10-01T10:54:54Z) - Adaptive Name Entity Recognition under Highly Unbalanced Data [5.575448433529451]
We present our experiments on a neural architecture composed of a Conditional Random Field (CRF) layer stacked on top of a Bi-directional LSTM (BI-LSTM) layer for solving NER tasks.
We introduce an add-on classification model to split sentences into two different sets: Weak and Strong classes and then designing a couple of Bi-LSTM-CRF models properly to optimize performance on each set.
arXiv Detail & Related papers (2020-03-10T06:56:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.