Modelling Verbal Morphology in Nen
- URL: http://arxiv.org/abs/2011.14489v2
- Date: Sun, 6 Dec 2020 23:08:01 GMT
- Title: Modelling Verbal Morphology in Nen
- Authors: Saliha Murado\u{g}lu, Nicholas Evans, Ekaterina Vylomova
- Abstract summary: We use state-of-the-art machine learning models for morphological reinflection to model Nen verbal morphology.
Our results show sensitivity to training data composition; different distributions of verb type yield different accuracies.
We also demonstrate the types of patterns that can be inferred from the training data through the case study of syncretism.
- Score: 4.6877729174041605
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Nen verbal morphology is remarkably complex; a transitive verb can take up to
1,740 unique forms. The combined effect of having a large combinatoric space
and a low-resource setting amplifies the need for NLP tools. Nen morphology
utilises distributed exponence - a non-trivial means of mapping form to
meaning. In this paper, we attempt to model Nen verbal morphology using
state-of-the-art machine learning models for morphological reinflection. We
explore and categorise the types of errors these systems generate. Our results
show sensitivity to training data composition; different distributions of verb
type yield different accuracies (patterning with E-complexity). We also
demonstrate the types of patterns that can be inferred from the training data
through the case study of syncretism.
Related papers
- Probabilistic Transformer: A Probabilistic Dependency Model for
Contextual Word Representation [52.270712965271656]
We propose a new model of contextual word representation, not from a neural perspective, but from a purely syntactic and probabilistic perspective.
We find that the graph of our model resembles transformers, with correspondences between dependencies and self-attention.
Experiments show that our model performs competitively to transformers on small to medium sized datasets.
arXiv Detail & Related papers (2023-11-26T06:56:02Z) - Morphological Inflection with Phonological Features [7.245355976804435]
This work explores effects on performance obtained through various ways in which morphological models get access to subcharacter phonological features.
We elicit phonemic data from standard graphemic data using language-specific grammars for languages with shallow grapheme-to-phoneme mapping.
arXiv Detail & Related papers (2023-06-21T21:34:39Z) - Dynamic Latent Separation for Deep Learning [67.62190501599176]
A core problem in machine learning is to learn expressive latent variables for model prediction on complex data.
Here, we develop an approach that improves expressiveness, provides partial interpretation, and is not restricted to specific applications.
arXiv Detail & Related papers (2022-10-07T17:56:53Z) - Modeling Target-Side Morphology in Neural Machine Translation: A
Comparison of Strategies [72.56158036639707]
Morphologically rich languages pose difficulties to machine translation.
A large amount of differently inflected word surface forms entails a larger vocabulary.
Some inflected forms of infrequent terms typically do not appear in the training corpus.
Linguistic agreement requires the system to correctly match the grammatical categories between inflected word forms in the output sentence.
arXiv Detail & Related papers (2022-03-25T10:13:20Z) - Tackling Morphological Analogies Using Deep Learning -- Extended Version [8.288496996031684]
Analogical proportions are statements of the form "A is to B as C is to D"
We propose an approach using deep learning to detect and solve morphological analogies.
We demonstrate our model's competitive performance on analogy detection and resolution over multiple languages.
arXiv Detail & Related papers (2021-11-09T13:45:23Z) - Model-agnostic multi-objective approach for the evolutionary discovery
of mathematical models [55.41644538483948]
In modern data science, it is more interesting to understand the properties of the model, which parts could be replaced to obtain better results.
We use multi-objective evolutionary optimization for composite data-driven model learning to obtain the algorithm's desired properties.
arXiv Detail & Related papers (2021-07-07T11:17:09Z) - Minimal Supervision for Morphological Inflection [8.532288965425805]
We bootstrapping labeled data from a seed as little as em five labeled paradigms, accompanied by a large bulk of unlabeled text.
Our approach exploits different kinds of regularities in morphological systems in a two-phased setup.
We experiment with the Paradigm Cell Filling Problem over eight typologically different languages, and find that, in languages with relatively simple morphology, orthographic regularities on their own allow inflection models to achieve respectable accuracy.
arXiv Detail & Related papers (2021-04-17T11:07:36Z) - Unnatural Language Inference [48.45003475966808]
We find that state-of-the-art NLI models, such as RoBERTa and BART, are invariant to, and sometimes even perform better on, examples with randomly reordered words.
Our findings call into question the idea that our natural language understanding models, and the tasks used for measuring their progress, genuinely require a human-like understanding of syntax.
arXiv Detail & Related papers (2020-12-30T20:40:48Z) - A Simple Joint Model for Improved Contextual Neural Lemmatization [60.802451210656805]
We present a simple joint neural model for lemmatization and morphological tagging that achieves state-of-the-art results on 20 languages.
Our paper describes the model in addition to training and decoding procedures.
arXiv Detail & Related papers (2019-04-04T02:03:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.