Review of Unsupervised POS Tagging and Its Implications on Language
Acquisition
- URL: http://arxiv.org/abs/2312.10169v1
- Date: Fri, 15 Dec 2023 19:31:00 GMT
- Title: Review of Unsupervised POS Tagging and Its Implications on Language
Acquisition
- Authors: Niels Dickson
- Abstract summary: An ability that underlies human syntactic knowledge is determining which words can appear in the similar structures.
In exploring this process, we will review various engineering approaches whose goal is similar to that of a child's.
We will discuss common themes that support the advances in the models and their relevance for language acquisition.
- Score: 0.0
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: An ability that underlies human syntactic knowledge is determining which
words can appear in the similar structures (i.e. grouping words by their
syntactic categories). These groupings enable humans to combine structures in
order to communicate complex meanings. A foundational question is how do
children acquire this ability underlying syntactic knowledge. In exploring this
process, we will review various engineering approaches whose goal is similar to
that of a child's -- without prior syntactic knowledge, correctly identify the
parts of speech (POS) of the words in a sample of text. In reviewing these
unsupervised tagging efforts, we will discuss common themes that support the
advances in the models and their relevance for language acquisition. For
example, we discuss how each model judges success (evaluation metrics), the
"additional information" that constrains the POS learning (such as orthographic
information), and the context used to determine POS (only previous word, words
before and after the target, etc). The identified themes pave the way for
future investigations into the cognitive processes that underpin the
acquisition of syntactic categories and provide a useful layout of current
state of the art unsupervised POS tagging models.
Related papers
- Human Inspired Progressive Alignment and Comparative Learning for
Grounded Word Acquisition [6.47452771256903]
We take inspiration from how human babies acquire their first language, and developed a computational process for word acquisition through comparative learning.
Motivated by cognitive findings, we generated a small dataset that enables the computation models to compare the similarities and differences of various attributes.
We frame the acquisition of words as not only the information filtration process, but also as representation-symbol mapping.
arXiv Detail & Related papers (2023-07-05T19:38:04Z) - Towards Open Vocabulary Learning: A Survey [146.90188069113213]
Deep neural networks have made impressive advancements in various core tasks like segmentation, tracking, and detection.
Recently, open vocabulary settings were proposed due to the rapid progress of vision language pre-training.
This paper provides a thorough review of open vocabulary learning, summarizing and analyzing recent developments in the field.
arXiv Detail & Related papers (2023-06-28T02:33:06Z) - Syntax and Semantics Meet in the "Middle": Probing the Syntax-Semantics
Interface of LMs Through Agentivity [68.8204255655161]
We present the semantic notion of agentivity as a case study for probing such interactions.
This suggests LMs may potentially serve as more useful tools for linguistic annotation, theory testing, and discovery.
arXiv Detail & Related papers (2023-05-29T16:24:01Z) - Prompting Language Models for Linguistic Structure [73.11488464916668]
We present a structured prompting approach for linguistic structured prediction tasks.
We evaluate this approach on part-of-speech tagging, named entity recognition, and sentence chunking.
We find that while PLMs contain significant prior knowledge of task labels due to task leakage into the pretraining corpus, structured prompting can also retrieve linguistic structure with arbitrary labels.
arXiv Detail & Related papers (2022-11-15T01:13:39Z) - A Linguistic Investigation of Machine Learning based Contradiction
Detection Models: An Empirical Analysis and Future Perspectives [0.34998703934432673]
We analyze two Natural Language Inference data sets with respect to their linguistic features.
The goal is to identify those syntactic and semantic properties that are particularly hard to comprehend for a machine learning model.
arXiv Detail & Related papers (2022-10-19T10:06:03Z) - Usage-based learning of grammatical categories [0.0]
This paper raises the question how these categories can be acquired and where they have come from.
We show that a categorial type network which has scores based on the success in a language interaction leads to the spontaneous formation of grammatical categories.
arXiv Detail & Related papers (2022-04-14T07:44:25Z) - Feature-rich multiplex lexical networks reveal mental strategies of
early language learning [0.7111443975103329]
We introduce FEature-Rich MUltiplex LEXical (FERMULEX) networks.
Similarities model heterogenous word associations across semantic/syntactic/phonological aspects of knowledge.
Words are enriched with multi-dimensional feature embeddings including frequency, age of acquisition, length and polysemy.
arXiv Detail & Related papers (2022-01-13T16:44:51Z) - Understanding Synonymous Referring Expressions via Contrastive Features [105.36814858748285]
We develop an end-to-end trainable framework to learn contrastive features on the image and object instance levels.
We conduct extensive experiments to evaluate the proposed algorithm on several benchmark datasets.
arXiv Detail & Related papers (2021-04-20T17:56:24Z) - Infusing Finetuning with Semantic Dependencies [62.37697048781823]
We show that, unlike syntax, semantics is not brought to the surface by today's pretrained models.
We then use convolutional graph encoders to explicitly incorporate semantic parses into task-specific finetuning.
arXiv Detail & Related papers (2020-12-10T01:27:24Z) - Introducing Syntactic Structures into Target Opinion Word Extraction
with Deep Learning [89.64620296557177]
We propose to incorporate the syntactic structures of the sentences into the deep learning models for targeted opinion word extraction.
We also introduce a novel regularization technique to improve the performance of the deep learning models.
The proposed model is extensively analyzed and achieves the state-of-the-art performance on four benchmark datasets.
arXiv Detail & Related papers (2020-10-26T07:13:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.