From Ultra-Fine to Fine: Fine-tuning Ultra-Fine Entity Typing Models to
Fine-grained
- URL: http://arxiv.org/abs/2312.06188v1
- Date: Mon, 11 Dec 2023 08:12:01 GMT
- Title: From Ultra-Fine to Fine: Fine-tuning Ultra-Fine Entity Typing Models to
Fine-grained
- Authors: Hongliang Dai, Ziqian Zeng
- Abstract summary: A common way to address this problem is to use distantly annotated training data that contains incorrect labels.
We propose a new approach that can avoid the need of creating distantly labeled data whenever there is a new type schema.
- Score: 12.948753628039093
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: For the task of fine-grained entity typing (FET), due to the use of a large
number of entity types, it is usually considered too costly to manually
annotating a training dataset that contains an ample number of examples for
each type. A common way to address this problem is to use distantly annotated
training data that contains incorrect labels. However, the performance of
models trained solely with such data can be limited by the errors in the
automatic annotation. Recently, there are a few approaches that no longer
follow this conventional way. But without using sufficient direct entity typing
supervision may also cause them to yield inferior performance. In this paper,
we propose a new approach that can avoid the need of creating distantly labeled
data whenever there is a new type schema. We first train an entity typing model
that have an extremely board type coverage by using the ultra-fine entity
typing data. Then, when there is a need to produce a model for a newly designed
fine-grained entity type schema. We can simply fine-tune the previously trained
model with a small number of examples annotated under this schema. Experimental
results show that our approach achieves outstanding performance for FET under
the few-shot setting. It can also outperform state-of-the-art weak supervision
based methods after fine-tuning the model with only a small size manually
annotated training set.
Related papers
- Seed-Guided Fine-Grained Entity Typing in Science and Engineering
Domains [51.02035914828596]
We study the task of seed-guided fine-grained entity typing in science and engineering domains.
We propose SEType which first enriches the weak supervision by finding more entities for each seen type from an unlabeled corpus.
It then matches the enriched entities to unlabeled text to get pseudo-labeled samples and trains a textual entailment model that can make inferences for both seen and unseen types.
arXiv Detail & Related papers (2024-01-23T22:36:03Z) - Calibrated Seq2seq Models for Efficient and Generalizable Ultra-fine
Entity Typing [10.08153231108538]
We present CASENT, a seq2seq model designed for ultra-fine entity typing.
Our model takes an entity mention as input and employs constrained beam search to generate multiple types autoregressively.
Our method outperforms the previous state-of-the-art in terms of F1 score and calibration error, while achieving an inference speedup of over 50 times.
arXiv Detail & Related papers (2023-11-01T20:39:12Z) - Taxonomy Expansion for Named Entity Recognition [65.49344005894996]
Training a Named Entity Recognition (NER) model often involves fixing a taxonomy of entity types.
A simple approach is to re-annotate entire dataset with both existing and additional entity types.
We propose a novel approach called Partial Label Model (PLM) that uses only partially annotated datasets.
arXiv Detail & Related papers (2023-05-22T16:23:46Z) - Ground Truth Inference for Weakly Supervised Entity Matching [76.6732856489872]
We propose a simple but powerful labeling model for weak supervision tasks.
We then tailor the labeling model specifically to the task of entity matching.
We show that our labeling model results in a 9% higher F1 score on average than the best existing method.
arXiv Detail & Related papers (2022-11-13T17:57:07Z) - Generative Entity Typing with Curriculum Learning [18.43562065432877]
We propose a novel generative entity typing (GET) paradigm.
Given a text with an entity mention, the multiple types for the role that the entity plays in the text are generated with a pre-trained language model.
Our experiments justify the superiority of our GET model over the state-of-the-art entity typing models.
arXiv Detail & Related papers (2022-10-06T13:32:50Z) - Few-Shot Fine-Grained Entity Typing with Automatic Label Interpretation
and Instance Generation [36.541309948222306]
We study the problem of few-shot Fine-grained Entity Typing (FET), where only a few annotated entity mentions with contexts are given for each entity type.
We propose a novel framework for few-shot FET consisting of two modules: (1) an entity type label interpretation module automatically learns to relate type labels to the vocabulary by jointly leveraging few-shot instances and the label hierarchy, and (2) a type-based contextualized instance generator produces new instances based on given instances to enlarge the training set for better generalization.
arXiv Detail & Related papers (2022-06-28T04:05:40Z) - Does Recommend-Revise Produce Reliable Annotations? An Analysis on
Missing Instances in DocRED [60.39125850987604]
We show that a textit-revise scheme results in false negative samples and an obvious bias towards popular entities and relations.
The relabeled dataset is released to serve as a more reliable test set of document RE models.
arXiv Detail & Related papers (2022-04-17T11:29:01Z) - Fast Model Editing at Scale [77.69220974621425]
We propose Model Editor Networks with Gradient Decomposition (MEND)
MEND is a collection of small auxiliary editing networks that use a single desired input-output pair to make fast, local edits to a pre-trained model.
MEND can be trained on a single GPU in less than a day even for 10 billion+ parameter models.
arXiv Detail & Related papers (2021-10-21T17:41:56Z) - Ultra-Fine Entity Typing with Weak Supervision from a Masked Language
Model [39.031515304057585]
Recently there is an effort to extend fine-grained entity typing by using a richer and ultra-fine set of types.
We propose to obtain training data for ultra-fine entity typing by using a BERT Masked Language Model (MLM)
Given a mention in a sentence, our approach constructs an input for the BERT so that it predicts context dependent hypernyms of the mention, which can be used as type labels.
arXiv Detail & Related papers (2021-06-08T04:43:28Z) - Interpretable Entity Representations through Large-Scale Typing [61.4277527871572]
We present an approach to creating entity representations that are human readable and achieve high performance out of the box.
Our representations are vectors whose values correspond to posterior probabilities over fine-grained entity types.
We show that it is possible to reduce the size of our type set in a learning-based way for particular domains.
arXiv Detail & Related papers (2020-04-30T23:58:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.