Tha3aroon at NSURL-2019 Task 8: Semantic Question Similarity in Arabic
- URL: http://arxiv.org/abs/1912.12514v1
- Date: Sat, 28 Dec 2019 20:11:33 GMT
- Title: Tha3aroon at NSURL-2019 Task 8: Semantic Question Similarity in Arabic
- Authors: Ali Fadel, Ibraheem Tuffaha, Mahmoud Al-Ayyoub
- Abstract summary: We describe our team's effort on the semantic text question similarity task of NSURL 2019.
Our top performing system utilizes several innovative data augmentation techniques to enlarge the training data.
It takes ELMo pre-trained contextual embeddings of the data and feeds them into an ON-LSTM network with self-attention.
- Score: 5.214494546503266
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we describe our team's effort on the semantic text question
similarity task of NSURL 2019. Our top performing system utilizes several
innovative data augmentation techniques to enlarge the training data. Then, it
takes ELMo pre-trained contextual embeddings of the data and feeds them into an
ON-LSTM network with self-attention. This results in sequence representation
vectors that are used to predict the relation between the question pairs. The
model is ranked in the 1st place with 96.499 F1-score (same as the second place
F1-score) and the 2nd place with 94.848 F1-score (differs by 1.076 F1-score
from the first place) on the public and private leaderboards, respectively.
Related papers
- Coreference Resolution through a seq2seq Transition-Based System [10.187353923159613]
We present a coreference resolution system that uses a text-to-text paradigm to predict mentions and links jointly.
We obtain state-of-the-art accuracy on the CoNLL-2012 datasets with 83.3 F1-score for English.
We get substantially higher zero-shot F1-scores for 3 out of 4 languages than previous approaches.
arXiv Detail & Related papers (2022-11-22T10:17:50Z) - Alibaba-Translate China's Submission for WMT 2022 Quality Estimation
Shared Task [80.22825549235556]
We present our submission to the sentence-level MQM benchmark at Quality Estimation Shared Task, named UniTE.
Specifically, our systems employ the framework of UniTE, which combined three types of input formats during training with a pre-trained language model.
Results show that our models reach 1st overall ranking in the Multilingual and English-Russian settings, and 2nd overall ranking in English-German and Chinese-English settings.
arXiv Detail & Related papers (2022-10-18T08:55:27Z) - Alibaba-Translate China's Submission for WMT 2022 Metrics Shared Task [61.34108034582074]
We build our system based on the core idea of UNITE (Unified Translation Evaluation)
During the model pre-training phase, we first apply the pseudo-labeled data examples to continuously pre-train UNITE.
During the fine-tuning phase, we use both Direct Assessment (DA) and Multidimensional Quality Metrics (MQM) data from past years' WMT competitions.
arXiv Detail & Related papers (2022-10-18T08:51:25Z) - Detecting Handwritten Mathematical Terms with Sensor Based Data [71.84852429039881]
We propose a solution to the UbiComp 2021 Challenge by Stabilo in which handwritten mathematical terms are supposed to be automatically classified.
The input data set contains data of different writers, with label strings constructed from a total of 15 different possible characters.
arXiv Detail & Related papers (2021-09-12T19:33:34Z) - SLADE: A Self-Training Framework For Distance Metric Learning [75.54078592084217]
We present a self-training framework, SLADE, to improve retrieval performance by leveraging additional unlabeled data.
We first train a teacher model on the labeled data and use it to generate pseudo labels for the unlabeled data.
We then train a student model on both labels and pseudo labels to generate final feature embeddings.
arXiv Detail & Related papers (2020-11-20T08:26:10Z) - NLP-CIC @ PRELEARN: Mastering prerequisites relations, from handcrafted
features to embeddings [68.97335984455059]
We present our systems and findings for the prerequisite relation learning task (PRELEARN) at EVALITA 2020.
The task aims to classify whether a pair of concepts hold a prerequisite relation or not.
Our submissions ranked first place in both scenarios with average F1 score of 0.887 and 0.690 respectively across domains on the test sets.
arXiv Detail & Related papers (2020-11-07T12:13:09Z) - Phonemer at WNUT-2020 Task 2: Sequence Classification Using COVID
Twitter BERT and Bagging Ensemble Technique based on Plurality Voting [0.0]
We develop a system that automatically identifies whether an English Tweet related to the novel coronavirus (COVID-19) is informative or not.
Our final approach achieved an F1-score of 0.9037 and we were ranked sixth overall with F1-score as the evaluation criteria.
arXiv Detail & Related papers (2020-10-01T10:54:54Z) - Relation Detection for Indonesian Language using Deep Neural Network --
Support Vector Machine [0.0]
We employ neural network to do relation detection between two named entities for Indonesian Language.
We used feature such as word embedding, position embedding, POS-Tag embedding, and character embedding.
The best result is 0.8083 on F1-Score using Convolutional Layer as front-part and SVM as back-part.
arXiv Detail & Related papers (2020-09-12T01:45:08Z) - Enhanced Universal Dependency Parsing with Second-Order Inference and
Mixture of Training Data [48.8386313914471]
This paper presents the system used in our submission to the textitIWPT 2020 Shared Task.
For the low-resource Tamil corpus, we specially mixed the training data of Tamil with other languages and significantly improved the performance of Tamil.
arXiv Detail & Related papers (2020-06-02T06:42:22Z) - The Inception Team at NSURL-2019 Task 8: Semantic Question Similarity in
Arabic [0.76146285961466]
This paper describes our method for the task of Semantic Question Similarity in Arabic.
The aim is to build a model that is able to detect similar semantic questions in the Arabic language for the provided dataset.
arXiv Detail & Related papers (2020-04-24T19:52:40Z) - UKARA 1.0 Challenge Track 1: Automatic Short-Answer Scoring in Bahasa
Indonesia [0.0]
We describe our third-place solution to the UKARA 1.0 challenge on automated essay scoring.
The task consists of a binary classification problem on two datasets | answers from two different questions.
We ended up using two different models for the two datasets.
arXiv Detail & Related papers (2020-02-28T04:32:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.