Knowledge-Enhanced Multi-Label Few-Shot Product Attribute-Value
Extraction
- URL: http://arxiv.org/abs/2308.08413v1
- Date: Wed, 16 Aug 2023 14:58:12 GMT
- Title: Knowledge-Enhanced Multi-Label Few-Shot Product Attribute-Value
Extraction
- Authors: Jiaying Gong, Wei-Te Chen, Hoda Eldardiry
- Abstract summary: Existing attribute-value extraction models require large quantities of labeled data for training.
New products with new attribute-value pairs enter the market every day in real-world e-Commerce.
We propose a Knowledge-Enhanced Attentive Framework (KEAF) based on networks to learn more discriminative prototypes.
- Score: 4.511923587827302
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Existing attribute-value extraction (AVE) models require large quantities of
labeled data for training. However, new products with new attribute-value pairs
enter the market every day in real-world e-Commerce. Thus, we formulate AVE in
multi-label few-shot learning (FSL), aiming to extract unseen attribute value
pairs based on a small number of training examples. We propose a
Knowledge-Enhanced Attentive Framework (KEAF) based on prototypical networks,
leveraging the generated label description and category information to learn
more discriminative prototypes. Besides, KEAF integrates with hybrid attention
to reduce noise and capture more informative semantics for each class by
calculating the label-relevant and query-related weights. To achieve
multi-label inference, KEAF further learns a dynamic threshold by integrating
the semantic information from both the support set and the query set. Extensive
experiments with ablation studies conducted on two datasets demonstrate that
KEAF outperforms other SOTA models for information extraction in FSL. The code
can be found at: https://github.com/gjiaying/KEAF
Related papers
- EIVEN: Efficient Implicit Attribute Value Extraction using Multimodal LLM [52.016009472409166]
EIVEN is a data- and parameter-efficient generative framework for implicit attribute value extraction.
We introduce a novel Learning-by-Comparison technique to reduce model confusion.
Our experiments reveal that EIVEN significantly outperforms existing methods in extracting implicit attribute values.
arXiv Detail & Related papers (2024-04-13T03:15:56Z) - Enhanced E-Commerce Attribute Extraction: Innovating with Decorative
Relation Correction and LLAMA 2.0-Based Annotation [4.81846973621209]
We propose a pioneering framework that integrates BERT for classification, a Conditional Random Fields (CRFs) layer for attribute value extraction, and Large Language Models (LLMs) for data annotation.
Our approach capitalizes on the robust representation learning of BERT, synergized with the sequence decoding prowess of CRFs, to adeptly identify and extract attribute values.
Our methodology is rigorously validated on various datasets, including Walmart, BestBuy's e-commerce NER dataset, and the CoNLL dataset.
arXiv Detail & Related papers (2023-12-09T08:26:30Z) - JPAVE: A Generation and Classification-based Model for Joint Product
Attribute Prediction and Value Extraction [59.94977231327573]
We propose a multi-task learning model with value generation/classification and attribute prediction called JPAVE.
Two variants of our model are designed for open-world and closed-world scenarios.
Experimental results on a public dataset demonstrate the superiority of our model compared with strong baselines.
arXiv Detail & Related papers (2023-11-07T18:36:16Z) - ActiveLab: Active Learning with Re-Labeling by Multiple Annotators [19.84626033109009]
ActiveLab is a method to decide what to label next in batch active learning.
It automatically estimates when it is more informative to re-label examples vs. labeling entirely new ones.
It reliably trains more accurate classifiers with far fewer annotations than a wide variety of popular active learning methods.
arXiv Detail & Related papers (2023-01-27T17:00:11Z) - Trustable Co-label Learning from Multiple Noisy Annotators [68.59187658490804]
Supervised deep learning depends on massive accurately annotated examples.
A typical alternative is learning from multiple noisy annotators.
This paper proposes a data-efficient approach, called emphTrustable Co-label Learning (TCL)
arXiv Detail & Related papers (2022-03-08T16:57:00Z) - QUEACO: Borrowing Treasures from Weakly-labeled Behavior Data for Query
Attribute Value Extraction [57.56700153507383]
This paper proposes a unified query attribute value extraction system in e-commerce search named QUEACO.
For the NER phase, QUEACO adopts a novel teacher-student network, where a teacher network that is trained on the strongly-labeled data generates pseudo-labels.
For the AVN phase, we also leverage the weakly-labeled query-to-attribute behavior data to normalize surface form attribute values from queries into canonical forms from products.
arXiv Detail & Related papers (2021-08-19T03:24:23Z) - Generate, Annotate, and Learn: Generative Models Advance Self-Training
and Knowledge Distillation [58.64720318755764]
Semi-Supervised Learning (SSL) has seen success in many application domains, but this success often hinges on the availability of task-specific unlabeled data.
Knowledge distillation (KD) has enabled compressing deep networks and ensembles, achieving the best results when distilling knowledge on fresh task-specific unlabeled examples.
We present a general framework called "generate, annotate, and learn (GAL)" that uses unconditional generative models to synthesize in-domain unlabeled data.
arXiv Detail & Related papers (2021-06-11T05:01:24Z) - Automatic Validation of Textual Attribute Values in E-commerce Catalog
by Learning with Limited Labeled Data [61.789797281676606]
We propose a novel meta-learning latent variable approach, called MetaBridge.
It can learn transferable knowledge from a subset of categories with limited labeled data.
It can capture the uncertainty of never-seen categories with unlabeled data.
arXiv Detail & Related papers (2020-06-15T21:31:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.