Parsing linearizations appreciate PoS tags - but some are fussy about
errors
- URL: http://arxiv.org/abs/2210.15219v1
- Date: Thu, 27 Oct 2022 07:15:36 GMT
- Title: Parsing linearizations appreciate PoS tags - but some are fussy about
errors
- Authors: Alberto Mu\~noz-Ortiz, Mark Anderson, David Vilares, Carlos
G\'omez-Rodr\'iguez
- Abstract summary: PoS tags, once taken for granted as a useful resource for syntactic parsing, have become more situational with the popularization of deep learning.
Recent work on the impact of PoS tags on graph- and transition-based labelings suggests that they are only useful when tagging accuracy is high, or in low-resource scenarios.
We undertake a study and uncover some trends. Among them, PoS tags are generally more useful for sequence labelings than for other paradigms, but the impact of their accuracy is highly encoding-dependent.
- Score: 12.024457689086008
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: PoS tags, once taken for granted as a useful resource for syntactic parsing,
have become more situational with the popularization of deep learning. Recent
work on the impact of PoS tags on graph- and transition-based parsers suggests
that they are only useful when tagging accuracy is prohibitively high, or in
low-resource scenarios. However, such an analysis is lacking for the emerging
sequence labeling parsing paradigm, where it is especially relevant as some
models explicitly use PoS tags for encoding and decoding. We undertake a study
and uncover some trends. Among them, PoS tags are generally more useful for
sequence labeling parsers than for other paradigms, but the impact of their
accuracy is highly encoding-dependent, with the PoS-based head-selection
encoding being best only when both tagging accuracy and resource availability
are high.
Related papers
- Weakly Supervised Deep Hyperspherical Quantization for Image Retrieval [76.4407063566172]
We propose Weakly-Supervised Deep Hyperspherical Quantization (WSDHQ), which is the first work to learn deep quantization from weakly tagged images.
Specifically, 1) we use word embeddings to represent the tags and enhance their semantic information based on a tag correlation graph.
We jointly learn semantics-preserving embeddings and supervised quantizer on hypersphere by employing a well-designed fusion layer and tailor-made loss functions.
arXiv Detail & Related papers (2024-04-07T15:48:33Z) - Another Dead End for Morphological Tags? Perturbed Inputs and Parsing [12.234169944475537]
We show that morphological tags can play a role to correct word-only neurals that make mistakes.
We also show that if morphological tags were utopically robust against lexical perturbations, they would be able to correct mistakes.
arXiv Detail & Related papers (2023-05-24T13:11:04Z) - Exploring Structured Semantic Prior for Multi Label Recognition with
Incomplete Labels [60.675714333081466]
Multi-label recognition (MLR) with incomplete labels is very challenging.
Recent works strive to explore the image-to-label correspondence in the vision-language model, ie, CLIP, to compensate for insufficient annotations.
We advocate remedying the deficiency of label supervision for the MLR with incomplete labels by deriving a structured semantic prior.
arXiv Detail & Related papers (2023-03-23T12:39:20Z) - On Parsing as Tagging [66.31276017088477]
We show how to reduce tetratagging, a state-of-the-art constituency tagger, to shift--reduce parsing.
We empirically evaluate our taxonomy of tagging pipelines with different choices of linearizers, learners, and decoders.
arXiv Detail & Related papers (2022-11-14T13:37:07Z) - Stylistic Fingerprints, POS-tags and Inflected Languages: A Case Study
in Polish [0.0]
Inflected languages make word forms sparse, making most statistical procedures complicated.
This paper examines the usefulness of grammatical features (as assessed via POS-tag n-grams) and lemmatized forms in recognizing author stylial profiles.
arXiv Detail & Related papers (2022-06-05T15:48:16Z) - On The Ingredients of an Effective Zero-shot Semantic Parser [95.01623036661468]
We analyze zero-shot learning by paraphrasing training examples of canonical utterances and programs from a grammar.
We propose bridging these gaps using improved grammars, stronger paraphrasers, and efficient learning methods.
Our model achieves strong performance on two semantic parsing benchmarks (Scholar, Geo) with zero labeled data.
arXiv Detail & Related papers (2021-10-15T21:41:16Z) - What Taggers Fail to Learn, Parsers Need the Most [0.38073142980733]
We present an error analysis of neural UPOS taggers to evaluate why using gold standard tags has such a large positive contribution to parsing performance.
We evaluate what neural dependencys implicitly learn about word types and how this relates to the errors taggers make to explain the minimal impact using predicted tags.
arXiv Detail & Related papers (2021-04-02T15:04:56Z) - PseudoSeg: Designing Pseudo Labels for Semantic Segmentation [78.35515004654553]
We present a re-design of pseudo-labeling to generate structured pseudo labels for training with unlabeled or weakly-labeled data.
We demonstrate the effectiveness of the proposed pseudo-labeling strategy in both low-data and high-data regimes.
arXiv Detail & Related papers (2020-10-19T17:59:30Z) - Fast semantic parsing with well-typedness guarantees [78.76675218975768]
AM dependency parsing is a principled method for neural semantic parsing with high accuracy across multiple graphbanks.
We describe an A* and a transition-based for AM dependency parsing which guarantee well-typedness and improve parsing speed by up to 3 orders of magnitude.
arXiv Detail & Related papers (2020-09-15T21:54:01Z) - Is POS Tagging Necessary or Even Helpful for Neural Dependency Parsing? [22.93722845643562]
We show that POS tagging can still significantly improve parsing performance when using the Stack joint framework.
Considering that it is much cheaper to annotate POS tags than parse trees, we also investigate the utilization of large-scale heterogeneous POS tag data.
arXiv Detail & Related papers (2020-03-06T13:47:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.