Easy-to-Hard Learning for Information Extraction
- URL: http://arxiv.org/abs/2305.09193v2
- Date: Fri, 19 May 2023 11:27:02 GMT
- Title: Easy-to-Hard Learning for Information Extraction
- Authors: Chang Gao, Wenxuan Zhang, Wai Lam, Lidong Bing
- Abstract summary: Information extraction systems aim to automatically extract structured information from unstructured texts.
We propose a unified easy-to-hard learning framework consisting of three stages, i.e., the easy stage, the hard stage, and the main stage.
By breaking down the learning process into multiple stages, our framework facilitates the model to acquire general IE task knowledge and improve its generalization ability.
- Score: 57.827955646831526
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Information extraction (IE) systems aim to automatically extract structured
information, such as named entities, relations between entities, and events,
from unstructured texts. While most existing work addresses a particular IE
task, universally modeling various IE tasks with one model has achieved great
success recently. Despite their success, they employ a one-stage learning
strategy, i.e., directly learning to extract the target structure given the
input text, which contradicts the human learning process. In this paper, we
propose a unified easy-to-hard learning framework consisting of three stages,
i.e., the easy stage, the hard stage, and the main stage, for IE by mimicking
the human learning process. By breaking down the learning process into multiple
stages, our framework facilitates the model to acquire general IE task
knowledge and improve its generalization ability. Extensive experiments across
four IE tasks demonstrate the effectiveness of our framework. We achieve new
state-of-the-art results on 13 out of 17 datasets. Our code is available at
\url{https://github.com/DAMO-NLP-SG/IE-E2H}.
Related papers
- RUIE: Retrieval-based Unified Information Extraction using Large Language Model [6.788855739199981]
Unified information extraction aims to complete all information extraction tasks using a single model or framework.
We propose RUIE (Retrieval-based Unified Information Extraction), a framework that leverages in-context learning to enable rapid generalization.
Experimental results on 8 held-out datasets demonstrate RUIE's effectiveness in generalizing to unseen tasks.
arXiv Detail & Related papers (2024-09-18T03:20:04Z) - A Regularization-based Transfer Learning Method for Information
Extraction via Instructed Graph Decoder [29.242560023747252]
We propose a regularization-based transfer learning method for IE (TIE) via an instructed graph decoder.
Specifically, we first construct an instruction pool for datasets from all well-known IE tasks, and then present an instructed graph decoder.
In this way, the common knowledge shared with existing datasets can be learned and transferred to a new dataset with new labels.
arXiv Detail & Related papers (2024-03-01T13:04:12Z) - Large Language Models for Generative Information Extraction: A Survey [89.71273968283616]
Large Language Models (LLMs) have demonstrated remarkable capabilities in text understanding and generation.
We present an extensive overview by categorizing these works in terms of various IE subtasks and techniques.
We empirically analyze the most advanced methods and discover the emerging trend of IE tasks with LLMs.
arXiv Detail & Related papers (2023-12-29T14:25:22Z) - Universal Information Extraction with Meta-Pretrained Self-Retrieval [39.69130086395689]
Universal Information Extraction(Universal IE) aims to solve different extraction tasks in a uniform text-to-structure generation manner.
Retrieving knowledge from external knowledge bases may help models to overcome this problem but it is impossible to construct a knowledge base suitable for various IE tasks.
We propose MetaRetriever to retrieve task-specific knowledge from PLMs to enhance universal IE.
arXiv Detail & Related papers (2023-06-18T00:16:00Z) - ICL-D3IE: In-Context Learning with Diverse Demonstrations Updating for
Document Information Extraction [56.790794611002106]
Large language models (LLMs) have demonstrated remarkable results in various natural language processing (NLP) tasks with in-context learning.
We propose a simple but effective in-context learning framework called ICL-D3IE.
Specifically, we extract the most difficult and distinct segments from hard training documents as hard demonstrations.
arXiv Detail & Related papers (2023-03-09T06:24:50Z) - Unified Structure Generation for Universal Information Extraction [58.89057387608414]
UIE can universally model different IE tasks, adaptively generate targeted structures, and collaboratively learn general IE abilities from different knowledge sources.
Experiments show that UIE achieved the state-of-the-art performance on 4 IE tasks, 13 datasets, and on all supervised, low-resource, and few-shot settings.
arXiv Detail & Related papers (2022-03-23T08:49:29Z) - Online Structured Meta-learning [137.48138166279313]
Current online meta-learning algorithms are limited to learn a globally-shared meta-learner.
We propose an online structured meta-learning (OSML) framework to overcome this limitation.
Experiments on three datasets demonstrate the effectiveness and interpretability of our proposed framework.
arXiv Detail & Related papers (2020-10-22T09:10:31Z) - Automated Relational Meta-learning [95.02216511235191]
We propose an automated relational meta-learning framework that automatically extracts the cross-task relations and constructs the meta-knowledge graph.
We conduct extensive experiments on 2D toy regression and few-shot image classification and the results demonstrate the superiority of ARML over state-of-the-art baselines.
arXiv Detail & Related papers (2020-01-03T07:02:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.