A Tale of a Probe and a Parser
- URL: http://arxiv.org/abs/2005.01641v2
- Date: Tue, 12 May 2020 10:38:43 GMT
- Title: A Tale of a Probe and a Parser
- Authors: Rowan Hall Maudslay, Josef Valvoda, Tiago Pimentel, Adina Williams,
Ryan Cotterell
- Abstract summary: Measuring what linguistic information is encoded in neural models of language has become popular in NLP.
Researchers approach this enterprise by training "probes" - supervised models designed to extract linguistic structure from another model's output.
One such probe is the structural probe, designed to quantify the extent to which syntactic information is encoded in contextualised word representations.
- Score: 74.14046092181947
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Measuring what linguistic information is encoded in neural models of language
has become popular in NLP. Researchers approach this enterprise by training
"probes" - supervised models designed to extract linguistic structure from
another model's output. One such probe is the structural probe (Hewitt and
Manning, 2019), designed to quantify the extent to which syntactic information
is encoded in contextualised word representations. The structural probe has a
novel design, unattested in the parsing literature, the precise benefit of
which is not immediately obvious. To explore whether syntactic probes would do
better to make use of existing techniques, we compare the structural probe to a
more traditional parser with an identical lightweight parameterisation. The
parser outperforms structural probe on UUAS in seven of nine analysed
languages, often by a substantial amount (e.g. by 11.1 points in English).
Under a second less common metric, however, there is the opposite trend - the
structural probe outperforms the parser. This begs the question: which metric
should we prefer?
Related papers
- Syntactic Language Change in English and German: Metrics, Parsers, and Convergences [56.47832275431858]
The current paper looks at diachronic trends in syntactic language change in both English and German, using corpora of parliamentary debates from the last c. 160 years.
We base our observations on five dependencys, including the widely used Stanford Core as well as 4 newer alternatives.
We show that changes in syntactic measures seem to be more frequent at the tails of sentence length distributions.
arXiv Detail & Related papers (2024-02-18T11:46:16Z) - Deep Lexical Hypothesis: Identifying personality structure in natural
language [0.30458514384586394]
We introduce a method to extract adjective similarities from language models.
The correlational structure produced through this method is highly similar to that of self- and other-ratings of 435 terms reported by Saucier and Goldberg.
Notably, Neuroticism and Openness are only weakly and inconsistently recovered.
arXiv Detail & Related papers (2022-03-04T02:06:10Z) - Do Syntactic Probes Probe Syntax? Experiments with Jabberwocky Probing [45.834234634602566]
We show that semantic cues in training data means that syntactic probes do not properly isolate syntax.
We train the probes on several popular language models.
arXiv Detail & Related papers (2021-06-04T15:46:39Z) - Intrinsic Probing through Dimension Selection [69.52439198455438]
Most modern NLP systems make use of pre-trained contextual representations that attain astonishingly high performance on a variety of tasks.
Such high performance should not be possible unless some form of linguistic structure inheres in these representations, and a wealth of research has sprung up on probing for it.
In this paper, we draw a distinction between intrinsic probing, which examines how linguistic information is structured within a representation, and the extrinsic probing popular in prior work, which only argues for the presence of such information by showing that it can be successfully extracted.
arXiv Detail & Related papers (2020-10-06T15:21:08Z) - Pareto Probing: Trading Off Accuracy for Complexity [87.09294772742737]
We argue for a probe metric that reflects the fundamental trade-off between probe complexity and performance.
Our experiments with dependency parsing reveal a wide gap in syntactic knowledge between contextual and non-contextual representations.
arXiv Detail & Related papers (2020-10-05T17:27:31Z) - Information-Theoretic Probing for Linguistic Structure [74.04862204427944]
We propose an information-theoretic operationalization of probing as estimating mutual information.
We evaluate on a set of ten typologically diverse languages often underrepresented in NLP research.
arXiv Detail & Related papers (2020-04-07T01:06:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.