STEAM: Self-Supervised Taxonomy Expansion with Mini-Paths
- URL: http://arxiv.org/abs/2006.10217v1
- Date: Thu, 18 Jun 2020 00:32:53 GMT
- Title: STEAM: Self-Supervised Taxonomy Expansion with Mini-Paths
- Authors: Yue Yu, Yinghao Li, Jiaming Shen, Hao Feng, Jimeng Sun and Chao Zhang
- Abstract summary: We propose a self-supervised taxonomy expansion model named STEAM.
STEAM generates natural self-supervision signals, and formulates a node attachment prediction task.
Experiments show STEAM outperforms state-of-the-art methods for taxonomy expansion by 11.6% in accuracy and 7.0% in mean reciprocal rank.
- Score: 53.45704816829921
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Taxonomies are important knowledge ontologies that underpin numerous
applications on a daily basis, but many taxonomies used in practice suffer from
the low coverage issue. We study the taxonomy expansion problem, which aims to
expand existing taxonomies with new concept terms. We propose a self-supervised
taxonomy expansion model named STEAM, which leverages natural supervision in
the existing taxonomy for expansion. To generate natural self-supervision
signals, STEAM samples mini-paths from the existing taxonomy, and formulates a
node attachment prediction task between anchor mini-paths and query terms. To
solve the node attachment task, it learns feature representations for
query-anchor pairs from multiple views and performs multi-view co-training for
prediction. Extensive experiments show that STEAM outperforms state-of-the-art
methods for taxonomy expansion by 11.6\% in accuracy and 7.0\% in mean
reciprocal rank on three public benchmarks. The implementation of STEAM can be
found at \url{https://github.com/yueyu1030/STEAM}.
Related papers
- CodeTaxo: Enhancing Taxonomy Expansion with Limited Examples via Code Language Prompts [40.52605902842168]
textscCodeTaxo is a novel approach that leverages large language models through code language prompts to capture the taxonomic structure.
Experiments on five real-world benchmarks from different domains demonstrate that textscCodeTaxo consistently achieves superior performance across all evaluation metrics.
arXiv Detail & Related papers (2024-08-17T02:15:07Z) - Using Zero-shot Prompting in the Automatic Creation and Expansion of
Topic Taxonomies for Tagging Retail Banking Transactions [0.0]
This work presents an unsupervised method for constructing and expanding topic using instruction-based fine-tuned LLMs (Large Language Models)
To expand an existing taxonomy with new terms, we use zero-shot prompting to find out where to add new nodes.
We use the resulting tags to assign tags that characterize merchants from a retail bank dataset.
arXiv Detail & Related papers (2024-01-08T00:27:16Z) - Towards Visual Taxonomy Expansion [50.462998483087915]
We propose Visual Taxonomy Expansion (VTE), introducing visual features into the taxonomy expansion task.
We propose a textual hypernymy learning task and a visual prototype learning task to cluster textual and visual semantics.
Our method is evaluated on two datasets, where we obtain compelling results.
arXiv Detail & Related papers (2023-09-12T10:17:28Z) - Taxonomy Expansion for Named Entity Recognition [65.49344005894996]
Training a Named Entity Recognition (NER) model often involves fixing a taxonomy of entity types.
A simple approach is to re-annotate entire dataset with both existing and additional entity types.
We propose a novel approach called Partial Label Model (PLM) that uses only partially annotated datasets.
arXiv Detail & Related papers (2023-05-22T16:23:46Z) - Who Should Go First? A Self-Supervised Concept Sorting Model for
Improving Taxonomy Expansion [50.794640012673064]
As data and business scope grow in real applications, existing need to be expanded to incorporate new concepts.
Previous works on taxonomy expansion process the new concepts independently and simultaneously, ignoring the potential relationships among them and the appropriate order of inserting operations.
We propose TaxoOrder, a novel self-supervised framework that simultaneously discovers the local hypernym-hyponym structure among new concepts and decides the order of insertion.
arXiv Detail & Related papers (2021-04-08T11:00:43Z) - Taxonomy Completion via Triplet Matching Network [18.37146040410778]
We formulate a new task, "taxonomy completion", by discovering both the hypernym and hyponym concepts for a query.
We propose Triplet Matching Network (TMN), to find the appropriate hypernym, hyponym> pairs for a given query concept.
TMN achieves the best performance on both taxonomy completion task and the previous taxonomy expansion task, outperforming existing methods.
arXiv Detail & Related papers (2021-01-06T07:19:55Z) - Octet: Online Catalog Taxonomy Enrichment with Self-Supervision [67.26804972901952]
We present a self-supervised end-to-end framework, Octet for Online Catalog EnrichmenT.
We propose to train a sequence labeling model for term extraction and employ graph neural networks (GNNs) to capture the taxonomy structure.
Octet enriches an online catalog in production to 2 times larger in the open-world evaluation.
arXiv Detail & Related papers (2020-06-18T04:53:07Z) - TaxoExpan: Self-supervised Taxonomy Expansion with Position-Enhanced
Graph Neural Network [62.12557274257303]
Taxonomies consist of machine-interpretable semantics and provide valuable knowledge for many web applications.
We propose a novel self-supervised framework, named TaxoExpan, which automatically generates a set of query concept, anchor concept> pairs from the existing taxonomy as training data.
We develop two innovative techniques in TaxoExpan: (1) a position-enhanced graph neural network that encodes the local structure of an anchor concept in the existing taxonomy, and (2) a noise-robust training objective that enables the learned model to be insensitive to the label noise in the self-supervision data.
arXiv Detail & Related papers (2020-01-26T21:30:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.