Learning What You Need from What You Did: Product Taxonomy Expansion
with User Behaviors Supervision
- URL: http://arxiv.org/abs/2203.14921v1
- Date: Mon, 28 Mar 2022 17:17:50 GMT
- Title: Learning What You Need from What You Did: Product Taxonomy Expansion
with User Behaviors Supervision
- Authors: Sijie Cheng, Zhouhong Gu, Bang Liu, Rui Xie, Wei Wu and Yanghua Xiao
- Abstract summary: We present a self-supervised and user behavior-oriented product expansion framework to append new concepts into existing taxonomy.
Our framework extracts hyponymy relations that conform to users' intentions and cognition.
Our method enlarges the size of real-world product from 39,263 to 94,698 relations with 88% semantic precision.
- Score: 21.649258076884927
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Taxonomies have been widely used in various domains to underpin numerous
applications. Specially, product taxonomies serve an essential role in the
e-commerce domain for the recommendation, browsing, and query understanding.
However, taxonomies need to constantly capture the newly emerged terms or
concepts in e-commerce platforms to keep up-to-date, which is expensive and
labor-intensive if it relies on manual maintenance and updates. Therefore, we
target the taxonomy expansion task to attach new concepts to existing
taxonomies automatically. In this paper, we present a self-supervised and user
behavior-oriented product taxonomy expansion framework to append new concepts
into existing taxonomies. Our framework extracts hyponymy relations that
conform to users' intentions and cognition. Specifically, i) to fully exploit
user behavioral information, we extract candidate hyponymy relations that match
user interests from query-click concepts; ii) to enhance the semantic
information of new concepts and better detect hyponymy relations, we model
concepts and relations through both user-generated content and structural
information in existing taxonomies and user click logs, by leveraging
Pre-trained Language Models and Graph Neural Network combined with Contrastive
Learning; iii) to reduce the cost of dataset construction and overcome data
skews, we construct a high-quality and balanced training dataset from existing
taxonomy with no supervision. Extensive experiments on real-world product
taxonomies in Meituan Platform, a leading Chinese vertical e-commerce platform
to order take-out with more than 70 million daily active users, demonstrate the
superiority of our proposed framework over state-of-the-art methods. Notably,
our method enlarges the size of real-world product taxonomies from 39,263 to
94,698 relations with 88% precision.
Related papers
- CodeTaxo: Enhancing Taxonomy Expansion with Limited Examples via Code Language Prompts [40.52605902842168]
textscCodeTaxo is a novel approach that leverages large language models through code language prompts to capture the taxonomic structure.
Experiments on five real-world benchmarks from different domains demonstrate that textscCodeTaxo consistently achieves superior performance across all evaluation metrics.
arXiv Detail & Related papers (2024-08-17T02:15:07Z) - Using Zero-shot Prompting in the Automatic Creation and Expansion of
Topic Taxonomies for Tagging Retail Banking Transactions [0.0]
This work presents an unsupervised method for constructing and expanding topic using instruction-based fine-tuned LLMs (Large Language Models)
To expand an existing taxonomy with new terms, we use zero-shot prompting to find out where to add new nodes.
We use the resulting tags to assign tags that characterize merchants from a retail bank dataset.
arXiv Detail & Related papers (2024-01-08T00:27:16Z) - TMT-VIS: Taxonomy-aware Multi-dataset Joint Training for Video Instance Segmentation [48.75470418596875]
Training on large-scale datasets can boost the performance of video instance segmentation while the datasets for VIS are hard to scale up due to the high labor cost.
What we possess are numerous isolated filed-specific datasets, thus, it is appealing to jointly train models across the aggregation of datasets to enhance data volume and diversity.
We conduct extensive evaluations on four popular and challenging benchmarks, including YouTube-VIS 2019, YouTube-VIS 2021, OVIS, and UVO.
Our model shows significant improvement over the baseline solutions, and sets new state-of-the-art records on all benchmarks.
arXiv Detail & Related papers (2023-12-11T18:50:09Z) - Towards Visual Taxonomy Expansion [50.462998483087915]
We propose Visual Taxonomy Expansion (VTE), introducing visual features into the taxonomy expansion task.
We propose a textual hypernymy learning task and a visual prototype learning task to cluster textual and visual semantics.
Our method is evaluated on two datasets, where we obtain compelling results.
arXiv Detail & Related papers (2023-09-12T10:17:28Z) - Who Should Go First? A Self-Supervised Concept Sorting Model for
Improving Taxonomy Expansion [50.794640012673064]
As data and business scope grow in real applications, existing need to be expanded to incorporate new concepts.
Previous works on taxonomy expansion process the new concepts independently and simultaneously, ignoring the potential relationships among them and the appropriate order of inserting operations.
We propose TaxoOrder, a novel self-supervised framework that simultaneously discovers the local hypernym-hyponym structure among new concepts and decides the order of insertion.
arXiv Detail & Related papers (2021-04-08T11:00:43Z) - Octet: Online Catalog Taxonomy Enrichment with Self-Supervision [67.26804972901952]
We present a self-supervised end-to-end framework, Octet for Online Catalog EnrichmenT.
We propose to train a sequence labeling model for term extraction and employ graph neural networks (GNNs) to capture the taxonomy structure.
Octet enriches an online catalog in production to 2 times larger in the open-world evaluation.
arXiv Detail & Related papers (2020-06-18T04:53:07Z) - STEAM: Self-Supervised Taxonomy Expansion with Mini-Paths [53.45704816829921]
We propose a self-supervised taxonomy expansion model named STEAM.
STEAM generates natural self-supervision signals, and formulates a node attachment prediction task.
Experiments show STEAM outperforms state-of-the-art methods for taxonomy expansion by 11.6% in accuracy and 7.0% in mean reciprocal rank.
arXiv Detail & Related papers (2020-06-18T00:32:53Z) - TaxoExpan: Self-supervised Taxonomy Expansion with Position-Enhanced
Graph Neural Network [62.12557274257303]
Taxonomies consist of machine-interpretable semantics and provide valuable knowledge for many web applications.
We propose a novel self-supervised framework, named TaxoExpan, which automatically generates a set of query concept, anchor concept> pairs from the existing taxonomy as training data.
We develop two innovative techniques in TaxoExpan: (1) a position-enhanced graph neural network that encodes the local structure of an anchor concept in the existing taxonomy, and (2) a noise-robust training objective that enables the learned model to be insensitive to the label noise in the self-supervision data.
arXiv Detail & Related papers (2020-01-26T21:30:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.