Universal Information Extraction as Unified Semantic Matching
- URL: http://arxiv.org/abs/2301.03282v1
- Date: Mon, 9 Jan 2023 11:51:31 GMT
- Title: Universal Information Extraction as Unified Semantic Matching
- Authors: Jie Lou, Yaojie Lu, Dai Dai, Wei Jia, Hongyu Lin, Xianpei Han, Le Sun,
Hua Wu
- Abstract summary: We decouple information extraction into two abilities, structuring and conceptualizing, which are shared by different tasks and schemas.
Based on this paradigm, we propose to universally model various IE tasks with Unified Semantic Matching framework.
In this way, USM can jointly encode schema and input text, uniformly extract substructures in parallel, and controllably decode target structures on demand.
- Score: 54.19974454019611
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The challenge of information extraction (IE) lies in the diversity of label
schemas and the heterogeneity of structures. Traditional methods require
task-specific model design and rely heavily on expensive supervision, making
them difficult to generalize to new schemas. In this paper, we decouple IE into
two basic abilities, structuring and conceptualizing, which are shared by
different tasks and schemas. Based on this paradigm, we propose to universally
model various IE tasks with Unified Semantic Matching (USM) framework, which
introduces three unified token linking operations to model the abilities of
structuring and conceptualizing. In this way, USM can jointly encode schema and
input text, uniformly extract substructures in parallel, and controllably
decode target structures on demand. Empirical evaluation on 4 IE tasks shows
that the proposed method achieves state-of-the-art performance under the
supervised experiments and shows strong generalization ability in zero/few-shot
transfer settings.
Related papers
- HCVP: Leveraging Hierarchical Contrastive Visual Prompt for Domain
Generalization [69.33162366130887]
Domain Generalization (DG) endeavors to create machine learning models that excel in unseen scenarios by learning invariant features.
We introduce a novel method designed to supplement the model with domain-level and task-specific characteristics.
This approach aims to guide the model in more effectively separating invariant features from specific characteristics, thereby boosting the generalization.
arXiv Detail & Related papers (2024-01-18T04:23:21Z) - Unified Multi-modal Unsupervised Representation Learning for
Skeleton-based Action Understanding [62.70450216120704]
Unsupervised pre-training has shown great success in skeleton-based action understanding.
We propose a Unified Multimodal Unsupervised Representation Learning framework, called UmURL.
UmURL exploits an efficient early-fusion strategy to jointly encode the multi-modal features in a single-stream manner.
arXiv Detail & Related papers (2023-11-06T13:56:57Z) - Concept-Centric Transformers: Enhancing Model Interpretability through
Object-Centric Concept Learning within a Shared Global Workspace [1.6574413179773757]
Concept-Centric Transformers is a simple yet effective configuration of the shared global workspace for interpretability.
We show that our model achieves better classification accuracy than all baselines across all problems.
arXiv Detail & Related papers (2023-05-25T06:37:39Z) - RexUIE: A Recursive Method with Explicit Schema Instructor for Universal
Information Extraction [47.89362854989252]
Universal Information Extraction is an area of interest due to the challenges posed by varying targets, heterogeneous structures, and demand-specific schemas.
Previous works have only achieved limited success by unifying a few tasks, such as Named Entity Recognition (NER) and Relation Extraction (RE)
In this paper, we redefine the authentic UIE with a formal formulation that encompasses almost all extraction schemas.
arXiv Detail & Related papers (2023-04-28T11:28:56Z) - Understanding and Constructing Latent Modality Structures in Multi-modal
Representation Learning [53.68371566336254]
We argue that the key to better performance lies in meaningful latent modality structures instead of perfect modality alignment.
Specifically, we design 1) a deep feature separation loss for intra-modality regularization; 2) a Brownian-bridge loss for inter-modality regularization; and 3) a geometric consistency loss for both intra- and inter-modality regularization.
arXiv Detail & Related papers (2023-03-10T14:38:49Z) - Unified Structure Generation for Universal Information Extraction [58.89057387608414]
UIE can universally model different IE tasks, adaptively generate targeted structures, and collaboratively learn general IE abilities from different knowledge sources.
Experiments show that UIE achieved the state-of-the-art performance on 4 IE tasks, 13 datasets, and on all supervised, low-resource, and few-shot settings.
arXiv Detail & Related papers (2022-03-23T08:49:29Z) - Interpretable MTL from Heterogeneous Domains using Boosted Tree [8.095372074268685]
Multi-task learning (MTL) aims at improving the generalization performance of several related tasks.
In this paper, following the philosophy of boosted tree, we proposed a two-stage method.
Experiments on both benchmark and real-world datasets validate the effectiveness of the proposed method.
arXiv Detail & Related papers (2020-03-16T08:58:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.