Yseop at FinSim-3 Shared Task 2021: Specializing Financial Domain
Learning with Phrase Representations
- URL: http://arxiv.org/abs/2108.09485v1
- Date: Sat, 21 Aug 2021 10:53:12 GMT
- Title: Yseop at FinSim-3 Shared Task 2021: Specializing Financial Domain
Learning with Phrase Representations
- Authors: Hanna Abi Akl, Dominique Mariko, Hugues de Mazancourt
- Abstract summary: We present our approaches for the FinSim-3 Shared Task 2021: Learning Semantic Similarities for the Financial Domain.
The aim of this task is to correctly classify a list of given terms from the financial domain into the most relevant hypernym.
Our system ranks 2nd overall on both metrics, scoring 0.917 on Average Accuracy and 1.141 on Mean Rank.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we present our approaches for the FinSim-3 Shared Task 2021:
Learning Semantic Similarities for the Financial Domain. The aim of this shared
task is to correctly classify a list of given terms from the financial domain
into the most relevant hypernym (or top-level) concept in an external ontology.
For our system submission, we evaluate two methods: a Sentence-RoBERTa
(SRoBERTa) embeddings model pre-trained on a custom corpus, and a dual
word-sentence embeddings model that builds on the first method by improving the
proposed baseline word embeddings construction using the FastText model to
boost the classification performance. Our system ranks 2nd overall on both
metrics, scoring 0.917 on Average Accuracy and 1.141 on Mean Rank.
Related papers
- Co-guiding for Multi-intent Spoken Language Understanding [53.30511968323911]
We propose a novel model termed Co-guiding Net, which implements a two-stage framework achieving the mutual guidances between the two tasks.
For the first stage, we propose single-task supervised contrastive learning, and for the second stage, we propose co-guiding supervised contrastive learning.
Experiment results on multi-intent SLU show that our model outperforms existing models by a large margin.
arXiv Detail & Related papers (2023-11-22T08:06:22Z) - Ahead of the Text: Leveraging Entity Preposition for Financial Relation
Extraction [0.0]
In the context of the ACM KDF- SIGIR 2023 competition, we undertook an entity relation task on a dataset of financial entity relations called REFind.
We fine-tuned the transformer-based language model roberta-large for text classification by utilizing a labeled training set to predict the entity relations.
As a result of our methodology, we achieved the 1st place ranking on the competition's public leaderboard.
arXiv Detail & Related papers (2023-08-08T18:56:52Z) - Prompt Algebra for Task Composition [131.97623832435812]
We consider Visual Language Models with prompt tuning as our base classifier.
We propose constrained prompt tuning to improve performance of the composite classifier.
On UTZappos it improves classification accuracy over the best base model by 8.45% on average.
arXiv Detail & Related papers (2023-06-01T03:20:54Z) - CompoundPiece: Evaluating and Improving Decompounding Performance of
Language Models [77.45934004406283]
We systematically study decompounding, the task of splitting compound words into their constituents.
We introduce a dataset of 255k compound and non-compound words across 56 diverse languages obtained from Wiktionary.
We introduce a novel methodology to train dedicated models for decompounding.
arXiv Detail & Related papers (2023-05-23T16:32:27Z) - DICoE@FinSim-3: Financial Hypernym Detection using Augmented Terms and
Distance-based Features [2.6599014990168834]
We present the submission of team DICoE for FinSim-3, the 3rd Shared Task on Learning Semantic Similarities for the Financial Domain.
The task provides a set of terms in the financial domain and requires to classify them into the most relevant hypernym from a financial ontology.
Our best-performing submission ranked 4th on the task's leaderboard.
arXiv Detail & Related papers (2021-09-30T08:01:48Z) - Term Expansion and FinBERT fine-tuning for Hypernym and Synonym Ranking
of Financial Terms [0.0]
We present systems that attempt to solve Hypernym and synonym matching problem.
We designed these systems to participate in the FinSim-3, a shared task of FinNLP workshop at IJCAI-2021.
Our best performing model (Accuracy: 0.917, Rank: 1.156) was developed by fine-tuning SentenceBERT [Reimers et al., 2019] over an extended labelled set created using the hierarchy of labels present in FIBO.
arXiv Detail & Related papers (2021-07-29T06:17:44Z) - NLP-CIC @ PRELEARN: Mastering prerequisites relations, from handcrafted
features to embeddings [68.97335984455059]
We present our systems and findings for the prerequisite relation learning task (PRELEARN) at EVALITA 2020.
The task aims to classify whether a pair of concepts hold a prerequisite relation or not.
Our submissions ranked first place in both scenarios with average F1 score of 0.887 and 0.690 respectively across domains on the test sets.
arXiv Detail & Related papers (2020-11-07T12:13:09Z) - SLM: Learning a Discourse Language Representation with Sentence
Unshuffling [53.42814722621715]
We introduce Sentence-level Language Modeling, a new pre-training objective for learning a discourse language representation.
We show that this feature of our model improves the performance of the original BERT by large margins.
arXiv Detail & Related papers (2020-10-30T13:33:41Z) - BRUMS at SemEval-2020 Task 3: Contextualised Embeddings for Predicting
the (Graded) Effect of Context in Word Similarity [9.710464466895521]
This paper presents the team BRUMS submission to SemEval-2020 Task 3: Graded Word Similarity in Context.
The system utilise state-of-the-art contextualised word embeddings, which have some task-specific adaptations, including stacked embeddings and average embeddings.
Following the final rankings, our approach is ranked within the top 5 solutions of each language while preserving the 1st position of Finnish subtask 2.
arXiv Detail & Related papers (2020-10-13T10:25:18Z) - NEMO: Frequentist Inference Approach to Constrained Linguistic Typology
Feature Prediction in SIGTYP 2020 Shared Task [83.43738174234053]
We employ frequentist inference to represent correlations between typological features and use this representation to train simple multi-class estimators that predict individual features.
Our best configuration achieved the micro-averaged accuracy score of 0.66 on 149 test languages.
arXiv Detail & Related papers (2020-10-12T19:25:43Z) - IITK at the FinSim Task: Hypernym Detection in Financial Domain via
Context-Free and Contextualized Word Embeddings [2.515934533974176]
FinSim 2020 task is to classify financial terms into the most relevant hypernym (or top-level) concept in an external ontology.
We leverage both context-dependent and context-independent word embeddings in our analysis.
Our system ranks 1st based on both the metrics, i.e. mean rank and accuracy.
arXiv Detail & Related papers (2020-07-22T04:56:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.