Named Entity Recognition with Small Strongly Labeled and Large Weakly
Labeled Data
- URL: http://arxiv.org/abs/2106.08977v1
- Date: Wed, 16 Jun 2021 17:18:14 GMT
- Title: Named Entity Recognition with Small Strongly Labeled and Large Weakly
Labeled Data
- Authors: Haoming Jiang, Danqing Zhang, Tianyu Cao, Bing Yin, Tuo Zhao
- Abstract summary: Weak supervision has shown promising results in many natural language processing tasks, such as Named Entity Recognition (NER)
We propose a new multi-stage computational framework -- NEEDLE with three essential ingredients: weak label completion, noise-aware loss function, and final fine-tuning over the strongly labeled data.
We demonstrate that NEEDLE can effectively suppress the noise of the weak labels and outperforms existing methods.
- Score: 37.980010197914105
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Weak supervision has shown promising results in many natural language
processing tasks, such as Named Entity Recognition (NER). Existing work mainly
focuses on learning deep NER models only with weak supervision, i.e., without
any human annotation, and shows that by merely using weakly labeled data, one
can achieve good performance, though still underperforms fully supervised NER
with manually/strongly labeled data. In this paper, we consider a more
practical scenario, where we have both a small amount of strongly labeled data
and a large amount of weakly labeled data. Unfortunately, we observe that
weakly labeled data does not necessarily improve, or even deteriorate the model
performance (due to the extensive noise in the weak labels) when we train deep
NER models over a simple or weighted combination of the strongly labeled and
weakly labeled data. To address this issue, we propose a new multi-stage
computational framework -- NEEDLE with three essential ingredients: (1) weak
label completion, (2) noise-aware loss function, and (3) final fine-tuning over
the strongly labeled data. Through experiments on E-commerce query NER and
Biomedical NER, we demonstrate that NEEDLE can effectively suppress the noise
of the weak labels and outperforms existing methods. In particular, we achieve
new SOTA F1-scores on 3 Biomedical NER datasets: BC5CDR-chem 93.74,
BC5CDR-disease 90.69, NCBI-disease 92.28.
Related papers
- Re-Examine Distantly Supervised NER: A New Benchmark and a Simple
Approach [15.87963432758696]
We critically assess the efficacy of current DS-NER methodologies using a real-world benchmark dataset named QTL.
To tackle the prevalent issue of label noise, we introduce a simple yet effective approach, Curriculum-based Positive-Unlabeled Learning CuPUL.
Our empirical results highlight the capability of CuPUL to significantly reduce the impact of noisy labels and outperform existing methods.
arXiv Detail & Related papers (2024-02-22T20:07:02Z) - Learning under Label Noise through Few-Shot Human-in-the-Loop Refinement [37.4838454216137]
Few-Shot Human-in-the-Loop Refinement (FHLR) is a novel solution to address noisy label learning.
We show that FHLR achieves significantly better performance when learning from noisy labels.
Our work not only achieves better generalization in high-stakes health sensing benchmarks but also sheds light on how noise affects commonly-used models.
arXiv Detail & Related papers (2024-01-25T11:43:35Z) - Enhancing Label Consistency on Document-level Named Entity Recognition [19.249781091058605]
Named entity recognition (NER) is a fundamental part of extracting information from documents in biomedical applications.
We present our method, ConNER, which enhances the label dependency of modifier (e.g., adjectives and prepositions) to achieve higher label agreement.
The effectiveness of our method is demonstrated on four popular biomedical NER datasets.
arXiv Detail & Related papers (2022-10-24T04:45:17Z) - How to tackle an emerging topic? Combining strong and weak labels for
Covid news NER [90.90053968189156]
We introduce a novel COVID-19 news NER dataset (COVIDNEWS-NER)
We release 3000 entries of hand annotated strongly labelled sentences and 13000 auto-generated weakly labelled sentences.
We show the effectiveness of CONTROSTER on COVIDNEWS-NER while providing analysis on combining weak and strong labels for training.
arXiv Detail & Related papers (2022-09-29T21:33:02Z) - Label Noise-Resistant Mean Teaching for Weakly Supervised Fake News
Detection [93.6222609806278]
We propose a novel label noise-resistant mean teaching approach (LNMT) for weakly supervised fake news detection.
LNMT leverages unlabeled news and feedback comments of users to enlarge the amount of training data.
LNMT establishes a mean teacher framework equipped with label propagation and label reliability estimation.
arXiv Detail & Related papers (2022-06-10T16:01:58Z) - Debiased Pseudo Labeling in Self-Training [77.83549261035277]
Deep neural networks achieve remarkable performances on a wide range of tasks with the aid of large-scale labeled datasets.
To mitigate the requirement for labeled data, self-training is widely used in both academia and industry by pseudo labeling on readily-available unlabeled data.
We propose Debiased, in which the generation and utilization of pseudo labels are decoupled by two independent heads.
arXiv Detail & Related papers (2022-02-15T02:14:33Z) - Boosting Semi-Supervised Face Recognition with Noise Robustness [54.342992887966616]
This paper presents an effective solution to semi-supervised face recognition that is robust to the label noise aroused by the auto-labelling.
We develop a semi-supervised face recognition solution, named Noise Robust Learning-Labelling (NRoLL), which is based on the robust training ability empowered by GN.
arXiv Detail & Related papers (2021-05-10T14:43:11Z) - Temporal Calibrated Regularization for Robust Noisy Label Learning [60.90967240168525]
Deep neural networks (DNNs) exhibit great success on many tasks with the help of large-scale well annotated datasets.
However, labeling large-scale data can be very costly and error-prone so that it is difficult to guarantee the annotation quality.
We propose a Temporal Calibrated Regularization (TCR) in which we utilize the original labels and the predictions in the previous epoch together.
arXiv Detail & Related papers (2020-07-01T04:48:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.