IITK at the FinSim Task: Hypernym Detection in Financial Domain via
Context-Free and Contextualized Word Embeddings
- URL: http://arxiv.org/abs/2007.11201v1
- Date: Wed, 22 Jul 2020 04:56:23 GMT
- Title: IITK at the FinSim Task: Hypernym Detection in Financial Domain via
Context-Free and Contextualized Word Embeddings
- Authors: Vishal Keswani, Sakshi Singh, Ashutosh Modi
- Abstract summary: FinSim 2020 task is to classify financial terms into the most relevant hypernym (or top-level) concept in an external ontology.
We leverage both context-dependent and context-independent word embeddings in our analysis.
Our system ranks 1st based on both the metrics, i.e. mean rank and accuracy.
- Score: 2.515934533974176
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we present our approaches for the FinSim 2020 shared task on
"Learning Semantic Representations for the Financial Domain". The goal of this
task is to classify financial terms into the most relevant hypernym (or
top-level) concept in an external ontology. We leverage both context-dependent
and context-independent word embeddings in our analysis. Our systems deploy
Word2vec embeddings trained from scratch on the corpus (Financial Prospectus in
English) along with pre-trained BERT embeddings. We divide the test dataset
into two subsets based on a domain rule. For one subset, we use unsupervised
distance measures to classify the term. For the second subset, we use simple
supervised classifiers like Naive Bayes, on top of the embeddings, to arrive at
a final prediction. Finally, we combine both the results. Our system ranks 1st
based on both the metrics, i.e., mean rank and accuracy.
Related papers
- Unify word-level and span-level tasks: NJUNLP's Participation for the
WMT2023 Quality Estimation Shared Task [59.46906545506715]
We introduce the NJUNLP team to the WMT 2023 Quality Estimation (QE) shared task.
Our team submitted predictions for the English-German language pair on all two sub-tasks.
Our models achieved the best results in English-German for both word-level and fine-grained error span detection sub-tasks.
arXiv Detail & Related papers (2023-09-23T01:52:14Z) - Relational Sentence Embedding for Flexible Semantic Matching [86.21393054423355]
We present Sentence Embedding (RSE), a new paradigm to discover further the potential of sentence embeddings.
RSE is effective and flexible in modeling sentence relations and outperforms a series of state-of-the-art embedding methods.
arXiv Detail & Related papers (2022-12-17T05:25:17Z) - DICoE@FinSim-3: Financial Hypernym Detection using Augmented Terms and
Distance-based Features [2.6599014990168834]
We present the submission of team DICoE for FinSim-3, the 3rd Shared Task on Learning Semantic Similarities for the Financial Domain.
The task provides a set of terms in the financial domain and requires to classify them into the most relevant hypernym from a financial ontology.
Our best-performing submission ranked 4th on the task's leaderboard.
arXiv Detail & Related papers (2021-09-30T08:01:48Z) - Yseop at FinSim-3 Shared Task 2021: Specializing Financial Domain
Learning with Phrase Representations [0.0]
We present our approaches for the FinSim-3 Shared Task 2021: Learning Semantic Similarities for the Financial Domain.
The aim of this task is to correctly classify a list of given terms from the financial domain into the most relevant hypernym.
Our system ranks 2nd overall on both metrics, scoring 0.917 on Average Accuracy and 1.141 on Mean Rank.
arXiv Detail & Related papers (2021-08-21T10:53:12Z) - Term Expansion and FinBERT fine-tuning for Hypernym and Synonym Ranking
of Financial Terms [0.0]
We present systems that attempt to solve Hypernym and synonym matching problem.
We designed these systems to participate in the FinSim-3, a shared task of FinNLP workshop at IJCAI-2021.
Our best performing model (Accuracy: 0.917, Rank: 1.156) was developed by fine-tuning SentenceBERT [Reimers et al., 2019] over an extended labelled set created using the hierarchy of labels present in FIBO.
arXiv Detail & Related papers (2021-07-29T06:17:44Z) - JSI at the FinSim-2 task: Ontology-Augmented Financial Concept
Classification [2.2559617939136505]
Ontologies are increasingly used for machine reasoning over the last few years.
This paper presents a practical use of an ontology for a classification problem from the financial domain.
We propose a method that maps given concepts to the mentioned explanations and performs a graph search for the most relevant hypernyms.
arXiv Detail & Related papers (2021-06-17T03:56:15Z) - Contrastive Learning and Self-Training for Unsupervised Domain
Adaptation in Semantic Segmentation [71.77083272602525]
UDA attempts to provide efficient knowledge transfer from a labeled source domain to an unlabeled target domain.
We propose a contrastive learning approach that adapts category-wise centroids across domains.
We extend our method with self-training, where we use a memory-efficient temporal ensemble to generate consistent and reliable pseudo-labels.
arXiv Detail & Related papers (2021-05-05T11:55:53Z) - R$^2$-Net: Relation of Relation Learning Network for Sentence Semantic
Matching [58.72111690643359]
We propose a Relation of Relation Learning Network (R2-Net) for sentence semantic matching.
We first employ BERT to encode the input sentences from a global perspective.
Then a CNN-based encoder is designed to capture keywords and phrase information from a local perspective.
To fully leverage labels for better relation information extraction, we introduce a self-supervised relation of relation classification task.
arXiv Detail & Related papers (2020-12-16T13:11:30Z) - RUSSE'2020: Findings of the First Taxonomy Enrichment Task for the
Russian language [70.27072729280528]
This paper describes the results of the first shared task on taxonomy enrichment for the Russian language.
16 teams participated in the task demonstrating high results with more than half of them outperforming the provided baseline.
arXiv Detail & Related papers (2020-05-22T13:30:37Z) - Words aren't enough, their order matters: On the Robustness of Grounding
Visual Referring Expressions [87.33156149634392]
We critically examine RefCOg, a standard benchmark for visual referring expression recognition.
We show that 83.7% of test instances do not require reasoning on linguistic structure.
We propose two methods, one based on contrastive learning and the other based on multi-task learning, to increase the robustness of ViLBERT.
arXiv Detail & Related papers (2020-05-04T17:09:15Z) - UiO-UvA at SemEval-2020 Task 1: Contextualised Embeddings for Lexical
Semantic Change Detection [5.099262949886174]
This paper focuses on Subtask 2, ranking words by the degree of their semantic drift over time.
We find that the most effective algorithms rely on the cosine similarity between averaged token embeddings and the pairwise distances between token embeddings.
arXiv Detail & Related papers (2020-04-30T18:43:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.