Local Additivity Based Data Augmentation for Semi-supervised NER
- URL: http://arxiv.org/abs/2010.01677v1
- Date: Sun, 4 Oct 2020 20:46:26 GMT
- Title: Local Additivity Based Data Augmentation for Semi-supervised NER
- Authors: Jiaao Chen, Zhenghui Wang, Ran Tian, Zichao Yang, Diyi Yang
- Abstract summary: Named Entity Recognition (NER) is one of the first stages in deep language understanding.
Current NER models heavily rely on human-annotated data.
We propose a Local Additivity based Data Augmentation (LADA) method for semi-supervised NER.
- Score: 59.90773003737093
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Named Entity Recognition (NER) is one of the first stages in deep language
understanding yet current NER models heavily rely on human-annotated data. In
this work, to alleviate the dependence on labeled data, we propose a Local
Additivity based Data Augmentation (LADA) method for semi-supervised NER, in
which we create virtual samples by interpolating sequences close to each other.
Our approach has two variations: Intra-LADA and Inter-LADA, where Intra-LADA
performs interpolations among tokens within one sentence, and Inter-LADA
samples different sentences to interpolate. Through linear additions between
sampled training data, LADA creates an infinite amount of labeled data and
improves both entity and context learning. We further extend LADA to the
semi-supervised setting by designing a novel consistency loss for unlabeled
data. Experiments conducted on two NER benchmarks demonstrate the effectiveness
of our methods over several strong baselines. We have publicly released our
code at https://github.com/GT-SALT/LADA.
Related papers
- Downstream-Pretext Domain Knowledge Traceback for Active Learning [138.02530777915362]
We propose a downstream-pretext domain knowledge traceback (DOKT) method that traces the data interactions of downstream knowledge and pre-training guidance.
DOKT consists of a traceback diversity indicator and a domain-based uncertainty estimator.
Experiments conducted on ten datasets show that our model outperforms other state-of-the-art methods.
arXiv Detail & Related papers (2024-07-20T01:34:13Z) - VALERIAN: Invariant Feature Learning for IMU Sensor-based Human Activity
Recognition in the Wild [7.50015216403068]
VALERIAN is an invariant feature learning method for in-the-wild wearable sensor-based HAR.
By training a multi-task model with separate task-specific layers for each subject, VALERIAN allows noisy labels to be dealt with individually while benefiting from shared feature representation across subjects.
arXiv Detail & Related papers (2023-03-03T18:22:14Z) - Linking data separation, visual separation, and classifier performance
using pseudo-labeling by contrastive learning [125.99533416395765]
We argue that the performance of the final classifier depends on the data separation present in the latent space and visual separation present in the projection.
We demonstrate our results by the classification of five real-world challenging image datasets of human intestinal parasites with only 1% supervised samples.
arXiv Detail & Related papers (2023-02-06T10:01:38Z) - A Dual-Contrastive Framework for Low-Resource Cross-Lingual Named Entity
Recognition [5.030581940990434]
Cross-lingual Named Entity Recognition (NER) has recently become a research hotspot because it can alleviate the data-hungry problem for low-resource languages.
In this paper, we describe our novel dual-contrastive framework ConCNER for cross-lingual NER under the scenario of limited source-language labeled data.
arXiv Detail & Related papers (2022-04-02T07:59:13Z) - Multi-Attribute Balanced Sampling for Disentangled GAN Controls [0.0]
Various controls over the generated data can be extracted from the latent space of a pre-trained GAN.
We show that this approach outperforms state-of-the-art classifier-based methods while avoiding the need for disentanglement-enforcing post-processing.
arXiv Detail & Related papers (2021-10-28T08:44:13Z) - Unsupervised Domain Adaptive Learning via Synthetic Data for Person
Re-identification [101.1886788396803]
Person re-identification (re-ID) has gained more and more attention due to its widespread applications in video surveillance.
Unfortunately, the mainstream deep learning methods still need a large quantity of labeled data to train models.
In this paper, we develop a data collector to automatically generate synthetic re-ID samples in a computer game, and construct a data labeler to simultaneously annotate them.
arXiv Detail & Related papers (2021-09-12T15:51:41Z) - Trash to Treasure: Harvesting OOD Data with Cross-Modal Matching for
Open-Set Semi-Supervised Learning [101.28281124670647]
Open-set semi-supervised learning (open-set SSL) investigates a challenging but practical scenario where out-of-distribution (OOD) samples are contained in the unlabeled data.
We propose a novel training mechanism that could effectively exploit the presence of OOD data for enhanced feature learning.
Our approach substantially lifts the performance on open-set SSL and outperforms the state-of-the-art by a large margin.
arXiv Detail & Related papers (2021-08-12T09:14:44Z) - Exploiting Image Translations via Ensemble Self-Supervised Learning for
Unsupervised Domain Adaptation [0.0]
We introduce an unsupervised domain adaption (UDA) strategy that combines multiple image translations, ensemble learning and self-supervised learning in one coherent approach.
We focus on one of the standard tasks of UDA in which a semantic segmentation model is trained on labeled synthetic data together with unlabeled real-world data.
arXiv Detail & Related papers (2021-07-13T16:43:02Z) - BOND: BERT-Assisted Open-Domain Named Entity Recognition with Distant
Supervision [49.42215511723874]
We propose a new computational framework -- BOND -- to improve the prediction performance of NER models.
Specifically, we propose a two-stage training algorithm: In the first stage, we adapt the pre-trained language model to the NER tasks using the distant labels.
In the second stage, we drop the distant labels, and propose a self-training approach to further improve the model performance.
arXiv Detail & Related papers (2020-06-28T04:55:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.