Unsupervised Full Constituency Parsing with Neighboring Distribution
Divergence
- URL: http://arxiv.org/abs/2110.15931v1
- Date: Fri, 29 Oct 2021 17:27:34 GMT
- Title: Unsupervised Full Constituency Parsing with Neighboring Distribution
Divergence
- Authors: Letian Peng, Zuchao Li, Hai Zhao
- Abstract summary: We propose an unsupervised and training-free labeling procedure by exploiting the property of a recently introduced metric.
For implementation, we develop NDD into Dual POS-NDD and build "molds" to detect constituents and their labels in sentences.
We show that DP-NDD not only labels constituents precisely but also inducts more accurate unlabeled constituency trees than all previous unsupervised methods with simpler rules.
- Score: 48.69930912510414
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unsupervised constituency parsing has been explored much but is still far
from being solved. Conventional unsupervised constituency parser is only able
to capture the unlabeled structure of sentences. Towards unsupervised full
constituency parsing, we propose an unsupervised and training-free labeling
procedure by exploiting the property of a recently introduced metric,
Neighboring Distribution Divergence (NDD), which evaluates semantic similarity
between sentences before and after editions. For implementation, we develop NDD
into Dual POS-NDD (DP-NDD) and build "molds" to detect constituents and their
labels in sentences. We show that DP-NDD not only labels constituents precisely
but also inducts more accurate unlabeled constituency trees than all previous
unsupervised methods with simpler rules. With two frameworks for labeled
constituency trees inference, we set both the new state-of-the-art for
unlabeled F1 and strong baselines for labeled F1. In contrast with the
conventional predicting-and-evaluating scenario, our method acts as an
plausible example to inversely apply evaluating metrics for prediction.
Related papers
- A Debiased Nearest Neighbors Framework for Multi-Label Text Classification [13.30576550077694]
We introduce a DEbiased Nearest Neighbors (DENN) framework for Multi-Label Text Classification (MLTC)
To address embedding alignment bias, we propose a debiased contrastive learning strategy, enhancing neighbor consistency on label co-occurrence.
For confidence estimation bias, we present a debiased confidence estimation strategy, improving the adaptive combination of predictions from $k$NN and inductive binary classifications.
arXiv Detail & Related papers (2024-08-06T14:00:23Z) - Appeal: Allow Mislabeled Samples the Chance to be Rectified in Partial Label Learning [55.4510979153023]
In partial label learning (PLL), each instance is associated with a set of candidate labels among which only one is ground-truth.
To help these mislabeled samples "appeal," we propose the first appeal-based framework.
arXiv Detail & Related papers (2023-12-18T09:09:52Z) - Cascading and Direct Approaches to Unsupervised Constituency Parsing on
Spoken Sentences [67.37544997614646]
We present the first study on unsupervised spoken constituency parsing.
The goal is to determine the spoken sentences' hierarchical syntactic structure in the form of constituency parse trees.
We show that accurate segmentation alone may be sufficient to parse spoken sentences accurately.
arXiv Detail & Related papers (2023-03-15T17:57:22Z) - Unreliable Partial Label Learning with Recursive Separation [44.901941653899264]
Unreliable Partial Label Learning (UPLL) is proposed, in which the true label may not be in the candidate label set.
We propose a two-stage framework named Unreliable Partial Label Learning with Recursive Separation (UPLLRS)
Our method demonstrates state-of-the-art performance as evidenced by experimental results.
arXiv Detail & Related papers (2023-02-20T10:39:31Z) - Dist-PU: Positive-Unlabeled Learning from a Label Distribution
Perspective [89.5370481649529]
We propose a label distribution perspective for PU learning in this paper.
Motivated by this, we propose to pursue the label distribution consistency between predicted and ground-truth label distributions.
Experiments on three benchmark datasets validate the effectiveness of the proposed method.
arXiv Detail & Related papers (2022-12-06T07:38:29Z) - Pointwise Binary Classification with Pairwise Confidence Comparisons [97.79518780631457]
We propose pairwise comparison (Pcomp) classification, where we have only pairs of unlabeled data that we know one is more likely to be positive than the other.
We link Pcomp classification to noisy-label learning to develop a progressive URE and improve it by imposing consistency regularization.
arXiv Detail & Related papers (2020-10-05T09:23:58Z) - MatchGAN: A Self-Supervised Semi-Supervised Conditional Generative
Adversarial Network [51.84251358009803]
We present a novel self-supervised learning approach for conditional generative adversarial networks (GANs) under a semi-supervised setting.
We perform augmentation by randomly sampling sensible labels from the label space of the few labelled examples available.
Our method surpasses the baseline with only 20% of the labelled examples used to train the baseline.
arXiv Detail & Related papers (2020-06-11T17:14:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.