JSI at the FinSim-2 task: Ontology-Augmented Financial Concept
Classification
- URL: http://arxiv.org/abs/2106.09230v1
- Date: Thu, 17 Jun 2021 03:56:15 GMT
- Title: JSI at the FinSim-2 task: Ontology-Augmented Financial Concept
Classification
- Authors: Timen Stepi\v{s}nik Perdih, Senja Pollak, Bla\v{z} \v{Skrlj}
- Abstract summary: Ontologies are increasingly used for machine reasoning over the last few years.
This paper presents a practical use of an ontology for a classification problem from the financial domain.
We propose a method that maps given concepts to the mentioned explanations and performs a graph search for the most relevant hypernyms.
- Score: 2.2559617939136505
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Ontologies are increasingly used for machine reasoning over the last few
years. They can provide explanations of concepts or be used for concept
classification if there exists a mapping from the desired labels to the
relevant ontology. Another advantage of using ontologies is that they do not
need a learning process, meaning that we do not need the train data or time
before using them. This paper presents a practical use of an ontology for a
classification problem from the financial domain. It first transforms a given
ontology to a graph and proceeds with generalization with the aim to find
common semantic descriptions of the input sets of financial concepts.
We present a solution to the shared task on Learning Semantic Similarities
for the Financial Domain (FinSim-2 task). The task is to design a system that
can automatically classify concepts from the Financial domain into the most
relevant hypernym concept in an external ontology - the Financial Industry
Business Ontology. We propose a method that maps given concepts to the
mentioned ontology and performs a graph search for the most relevant hypernyms.
We also employ a word vectorization method and a machine learning classifier to
supplement the method with a ranked list of labels for each concept.
Related papers
- Discover-then-Name: Task-Agnostic Concept Bottlenecks via Automated Concept Discovery [52.498055901649025]
Concept Bottleneck Models (CBMs) have been proposed to address the 'black-box' problem of deep neural networks.
We propose a novel CBM approach -- called Discover-then-Name-CBM (DN-CBM) -- that inverts the typical paradigm.
Our concept extraction strategy is efficient, since it is agnostic to the downstream task, and uses concepts already known to the model.
arXiv Detail & Related papers (2024-07-19T17:50:11Z) - Ontology Embedding: A Survey of Methods, Applications and Resources [54.3453925775069]
Ontologies are widely used for representing domain knowledge and meta data.
One straightforward solution is to integrate statistical analysis and machine learning.
Numerous papers have been published on embedding, but a lack of systematic reviews hinders researchers from gaining a comprehensive understanding of this field.
arXiv Detail & Related papers (2024-06-16T14:49:19Z) - OntoMath${}^{\mathbf{PRO}}$ 2.0 Ontology: Updates of the Formal Model [68.8204255655161]
The main attention is paid to the development of a formal model for the representation of mathematical statements in the Open Linked Data cloud.
The proposed model is intended for applications that extract mathematical facts from natural language mathematical texts and represent these facts as Linked Open Data.
The model is used in development of a new version of the OntoMath$mathrmPRO$ ontology of professional mathematics is described.
arXiv Detail & Related papers (2023-03-17T20:29:17Z) - Semantic Search for Large Scale Clinical Ontologies [63.71950996116403]
We present a deep learning approach to build a search system for large clinical vocabularies.
We propose a Triplet-BERT model and a method that generates training data based on semantic training data.
The model is evaluated using five real benchmark data sets and the results show that our approach achieves high results on both free text to concept and concept to searching concept vocabularies.
arXiv Detail & Related papers (2022-01-01T05:15:42Z) - DICoE@FinSim-3: Financial Hypernym Detection using Augmented Terms and
Distance-based Features [2.6599014990168834]
We present the submission of team DICoE for FinSim-3, the 3rd Shared Task on Learning Semantic Similarities for the Financial Domain.
The task provides a set of terms in the financial domain and requires to classify them into the most relevant hypernym from a financial ontology.
Our best-performing submission ranked 4th on the task's leaderboard.
arXiv Detail & Related papers (2021-09-30T08:01:48Z) - Active Refinement for Multi-Label Learning: A Pseudo-Label Approach [84.52793080276048]
Multi-label learning (MLL) aims to associate a given instance with its relevant labels from a set of concepts.
Previous works of MLL mainly focused on the setting where the concept set is assumed to be fixed.
Many real-world applications require introducing new concepts into the set to meet new demands.
arXiv Detail & Related papers (2021-09-29T19:17:05Z) - FinMatcher at FinSim-2: Hypernym Detection in the Financial Services
Domain using Knowledge Graphs [1.2891210250935146]
This paper presents the FinMatcher system and its results for the FinSim 2021 shared task.
The FinSim-2 shared task consists of a set of concept labels from the financial services domain.
The goal is to find the most relevant top-level concept from a given set of concepts.
arXiv Detail & Related papers (2021-03-02T08:56:28Z) - IITK at the FinSim Task: Hypernym Detection in Financial Domain via
Context-Free and Contextualized Word Embeddings [2.515934533974176]
FinSim 2020 task is to classify financial terms into the most relevant hypernym (or top-level) concept in an external ontology.
We leverage both context-dependent and context-independent word embeddings in our analysis.
Our system ranks 1st based on both the metrics, i.e. mean rank and accuracy.
arXiv Detail & Related papers (2020-07-22T04:56:23Z) - R-VGAE: Relational-variational Graph Autoencoder for Unsupervised
Prerequisite Chain Learning [83.13634692459486]
We propose a model called Graph AutoEncoder (VGA-E) to predict concept relations within a graph consisting of concept resource nodes.
Results show that our unsupervised approach outperforms graph-based semi-supervised methods and other baseline methods by up to 9.77% and 10.47% in terms of prerequisite relation prediction accuracy and F1 score.
Our method is notably the first graph-based model that attempts to make use of deep learning representations for the task of unsupervised prerequisite learning.
arXiv Detail & Related papers (2020-04-22T14:48:03Z) - Distributional semantic modeling: a revised technique to train term/word
vector space models applying the ontology-related approach [36.248702416150124]
We design a new technique for the distributional semantic modeling with a neural network-based approach to learn distributed term representations (or term embeddings)
Vec2graph is a Python library for visualizing word embeddings (term embeddings in our case) as dynamic and interactive graphs.
arXiv Detail & Related papers (2020-03-06T18:27:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.