Knowledge-guided Open Attribute Value Extraction with Reinforcement
Learning
- URL: http://arxiv.org/abs/2010.09189v1
- Date: Mon, 19 Oct 2020 03:28:27 GMT
- Title: Knowledge-guided Open Attribute Value Extraction with Reinforcement
Learning
- Authors: Ye Liu, Sheng Zhang, Rui Song, Suo Feng, Yanghua Xiao
- Abstract summary: We propose a knowledge-guided reinforcement learning (RL) framework for open attribute value extraction.
We trained a deep Q-network to sequentially compare extracted answers to improve extraction accuracy.
Our results show that our method outperforms the baselines by 16.5 - 27.8%.
- Score: 23.125544502927482
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Open attribute value extraction for emerging entities is an important but
challenging task. A lot of previous works formulate the problem as a
\textit{question-answering} (QA) task. While the collections of articles from
web corpus provide updated information about the emerging entities, the
retrieved texts can be noisy, irrelevant, thus leading to inaccurate answers.
Effectively filtering out noisy articles as well as bad answers is the key to
improving extraction accuracy. Knowledge graph (KG), which contains rich, well
organized information about entities, provides a good resource to address the
challenge. In this work, we propose a knowledge-guided reinforcement learning
(RL) framework for open attribute value extraction. Informed by relevant
knowledge in KG, we trained a deep Q-network to sequentially compare extracted
answers to improve extraction accuracy. The proposed framework is applicable to
different information extraction system. Our experimental results show that our
method outperforms the baselines by 16.5 - 27.8\%.
Related papers
- Retriever-and-Memory: Towards Adaptive Note-Enhanced Retrieval-Augmented Generation [72.70046559930555]
We propose a generic RAG approach called Adaptive Note-Enhanced RAG (Adaptive-Note) for complex QA tasks.
Specifically, Adaptive-Note introduces an overarching view of knowledge growth, iteratively gathering new information in the form of notes.
In addition, we employ an adaptive, note-based stop-exploration strategy to decide "what to retrieve and when to stop" to encourage sufficient knowledge exploration.
arXiv Detail & Related papers (2024-10-11T14:03:29Z) - EWEK-QA: Enhanced Web and Efficient Knowledge Graph Retrieval for Citation-based Question Answering Systems [103.91826112815384]
citation-based QA systems are suffering from two shortcomings.
They usually rely only on web as a source of extracted knowledge and adding other external knowledge sources can hamper the efficiency of the system.
We propose our enhanced web and efficient knowledge graph (KG) retrieval solution (EWEK-QA) to enrich the content of the extracted knowledge fed to the system.
arXiv Detail & Related papers (2024-06-14T19:40:38Z) - Assessing the quality of information extraction [0.0]
We introduce an automatic framework to assess the quality of the information extraction/retrieval and its completeness.
We discuss how to handle the input/output size limitations of the large language models and analyze their performance.
arXiv Detail & Related papers (2024-04-05T12:51:48Z) - Revisiting Sparse Retrieval for Few-shot Entity Linking [33.15662306409253]
We propose an ELECTRA-based keyword extractor to denoise the mention context and construct a better query expression.
For training the extractor, we propose a distant supervision method to automatically generate training data based on overlapping tokens between mention contexts and entity descriptions.
Experimental results on the ZESHEL dataset demonstrate that the proposed method outperforms state-of-the-art models by a significant margin across all test domains.
arXiv Detail & Related papers (2023-10-19T03:51:10Z) - Enriching Relation Extraction with OpenIE [70.52564277675056]
Relation extraction (RE) is a sub-discipline of information extraction (IE)
In this work, we explore how recent approaches for open information extraction (OpenIE) may help to improve the task of RE.
Our experiments over two annotated corpora, KnowledgeNet and FewRel, demonstrate the improved accuracy of our enriched models.
arXiv Detail & Related papers (2022-12-19T11:26:23Z) - Knowledge-Enhanced Relation Extraction Dataset [8.612433805862619]
There is currently no public dataset that encompasses both evidence sentences and knowledge graphs for knowledge-enhanced relation extraction.
We introduce the Knowledge-Enhanced Relation Extraction dataset (KERED)
KERED annotates each sentence with a relational fact, and it provides knowledge context for entities through entity linking.
arXiv Detail & Related papers (2022-10-19T13:23:10Z) - A Unified End-to-End Retriever-Reader Framework for Knowledge-based VQA [67.75989848202343]
This paper presents a unified end-to-end retriever-reader framework towards knowledge-based VQA.
We shed light on the multi-modal implicit knowledge from vision-language pre-training models to mine its potential in knowledge reasoning.
Our scheme is able to not only provide guidance for knowledge retrieval, but also drop these instances potentially error-prone towards question answering.
arXiv Detail & Related papers (2022-06-30T02:35:04Z) - BertNet: Harvesting Knowledge Graphs with Arbitrary Relations from
Pretrained Language Models [65.51390418485207]
We propose a new approach of harvesting massive KGs of arbitrary relations from pretrained LMs.
With minimal input of a relation definition, the approach efficiently searches in the vast entity pair space to extract diverse accurate knowledge.
We deploy the approach to harvest KGs of over 400 new relations from different LMs.
arXiv Detail & Related papers (2022-06-28T19:46:29Z) - Enhanced Language Representation with Label Knowledge for Span
Extraction [2.4909170697740963]
We introduce a new paradigm to integrate label knowledge and propose a novel model to explicitly and efficiently integrate label knowledge into text representations.
Specifically, it encodes texts and label annotations independently and then integrates label knowledge into text representation with an elaborate-designed semantics fusion module.
We conduct extensive experiments on three typical span extraction tasks: flat NER, nested NER, and event detection.
Our method achieves state-of-the-art performance on four benchmarks, and 2) reduces training time and inference time by 76% and 77% on average, respectively, compared with the QA Formalization paradigm.
arXiv Detail & Related papers (2021-11-01T12:21:05Z) - BERTese: Learning to Speak to BERT [50.76152500085082]
We propose a method for automatically rewriting queries into "BERTese", a paraphrase query that is directly optimized towards better knowledge extraction.
We empirically show our approach outperforms competing baselines, obviating the need for complex pipelines.
arXiv Detail & Related papers (2021-03-09T10:17:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.