Focused Contrastive Training for Test-based Constituency Analysis
- URL: http://arxiv.org/abs/2109.15159v1
- Date: Thu, 30 Sep 2021 14:22:15 GMT
- Title: Focused Contrastive Training for Test-based Constituency Analysis
- Authors: Benjamin Roth, Erion \c{C}ano
- Abstract summary: We propose a scheme for self-training of grammaticality models for constituency analysis based on linguistic tests.
A pre-trained language model is fine-tuned by contrastive estimation of grammatical sentences from a corpus, and ungrammatical sentences that were perturbed by a syntactic test.
- Score: 7.312581661832785
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a scheme for self-training of grammaticality models for
constituency analysis based on linguistic tests. A pre-trained language model
is fine-tuned by contrastive estimation of grammatical sentences from a corpus,
and ungrammatical sentences that were perturbed by a syntactic test, a
transformation that is motivated by constituency theory. We show that
consistent gains can be achieved if only certain positive instances are chosen
for training, depending on whether they could be the result of a test
transformation. This way, the positives, and negatives exhibit similar
characteristics, which makes the objective more challenging for the language
model, and also allows for additional markup that indicates the position of the
test application within the sentence.
Related papers
- Distilling Monolingual and Crosslingual Word-in-Context Representations [18.87665111304974]
We propose a method that distils representations of word meaning in context from a pre-trained language model in both monolingual and crosslingual settings.
Our method does not require human-annotated corpora nor updates of the parameters of the pre-trained model.
Our method learns to combine the outputs of different hidden layers of the pre-trained model using self-attention.
arXiv Detail & Related papers (2024-09-13T11:10:16Z) - Disco-Bench: A Discourse-Aware Evaluation Benchmark for Language
Modelling [70.23876429382969]
We propose a benchmark that can evaluate intra-sentence discourse properties across a diverse set of NLP tasks.
Disco-Bench consists of 9 document-level testsets in the literature domain, which contain rich discourse phenomena.
For linguistic analysis, we also design a diagnostic test suite that can examine whether the target models learn discourse knowledge.
arXiv Detail & Related papers (2023-07-16T15:18:25Z) - ChatGPT as a Text Simplification Tool to Remove Bias [0.0]
The presence of specific linguistic signals particular to a certain sub-group can be picked up by language models during training.
We explore a potential technique for bias mitigation in the form of simplification of text.
arXiv Detail & Related papers (2023-05-09T13:10:23Z) - Not another Negation Benchmark: The NaN-NLI Test Suite for Sub-clausal
Negation [59.307534363825816]
Negation is poorly captured by current language models, although the extent of this problem is not widely understood.
We introduce a natural language inference (NLI) test suite to enable probing the capabilities of NLP methods.
arXiv Detail & Related papers (2022-10-06T23:39:01Z) - Discontinuous Constituency and BERT: A Case Study of Dutch [0.20305676256390934]
We quantify the syntactic capacity of BERT in the evaluation regime of non-context free patterns, as occurring in Dutch.
We devise a test suite based on a mildly context-sensitive formalism, from which we derive grammars that capture the linguistic phenomena of control verb nesting and verb raising.
Our results, backed by extensive analysis, suggest that the models investigated fail in the implicit acquisition of the dependencies examined.
arXiv Detail & Related papers (2022-03-02T12:30:21Z) - Self-Normalized Importance Sampling for Neural Language Modeling [97.96857871187052]
In this work, we propose self-normalized importance sampling. Compared to our previous work, the criteria considered in this work are self-normalized and there is no need to further conduct a correction step.
We show that our proposed self-normalized importance sampling is competitive in both research-oriented and production-oriented automatic speech recognition tasks.
arXiv Detail & Related papers (2021-11-11T16:57:53Z) - Empowering Language Understanding with Counterfactual Reasoning [141.48592718583245]
We propose a Counterfactual Reasoning Model, which mimics the counterfactual thinking by learning from few counterfactual samples.
In particular, we devise a generation module to generate representative counterfactual samples for each factual sample, and a retrospective module to retrospect the model prediction by comparing the counterfactual and factual samples.
arXiv Detail & Related papers (2021-06-06T06:36:52Z) - Refining Targeted Syntactic Evaluation of Language Models [6.991281327290524]
Targeted syntactic evaluation of subject-verb number agreement in English (TSE)
Method evaluates whether language models rate each grammatical sentence as more likely than its ungrammatical counterpart.
We find that TSE overestimates systematicity of language models, but that models score up to 40% better on verbs that they predict are likely in context.
arXiv Detail & Related papers (2021-04-19T20:55:13Z) - Unsupervised Parsing via Constituency Tests [49.42244463346612]
We propose a method for unsupervised parsing based on the linguistic notion of a constituency test.
To produce a tree given a sentence, we score each span by aggregating its constituency test judgments, and we choose the binary tree with the highest total score.
The refined model achieves 62.8 F1 on the Penn Treebank test set, an absolute improvement of 7.6 points over the previous best published result.
arXiv Detail & Related papers (2020-10-07T04:05:01Z) - Grounded Compositional Outputs for Adaptive Language Modeling [59.02706635250856]
A language model's vocabulary$-$typically selected before training and permanently fixed later$-$affects its size.
We propose a fully compositional output embedding layer for language models.
To our knowledge, the result is the first word-level language model with a size that does not depend on the training vocabulary.
arXiv Detail & Related papers (2020-09-24T07:21:14Z) - Recurrent Neural Network Language Models Always Learn English-Like
Relative Clause Attachment [17.995905582226463]
We compare model performance in English and Spanish to show that non-linguistic biases in RNN LMs advantageously overlap with syntactic structure in English but not Spanish.
English models may appear to acquire human-like syntactic preferences, while models trained on Spanish fail to acquire comparable human-like preferences.
arXiv Detail & Related papers (2020-05-01T01:21:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.