The IMS-CUBoulder System for the SIGMORPHON 2020 Shared Task on
Unsupervised Morphological Paradigm Completion
- URL: http://arxiv.org/abs/2005.12411v1
- Date: Mon, 25 May 2020 21:23:52 GMT
- Title: The IMS-CUBoulder System for the SIGMORPHON 2020 Shared Task on
Unsupervised Morphological Paradigm Completion
- Authors: Manuel Mager and Katharina Kann
- Abstract summary: We present the systems of the University of Stuttgart IMS and the University of Colorado Boulder for SIGMORPHON 2020 Task 2 on unsupervised morphological paradigm completion.
The task consists of generating the morphological paradigms of a set of lemmas, given only the lemmas themselves and unlabeled text.
Our pointer-generator system obtains the best score of all seven submitted systems on average over all languages, and outperforms the official baseline, which was best overall, on Bulgarian and Kannada.
- Score: 27.37360427124081
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: In this paper, we present the systems of the University of Stuttgart IMS and
the University of Colorado Boulder (IMS-CUBoulder) for SIGMORPHON 2020 Task 2
on unsupervised morphological paradigm completion (Kann et al., 2020). The task
consists of generating the morphological paradigms of a set of lemmas, given
only the lemmas themselves and unlabeled text. Our proposed system is a
modified version of the baseline introduced together with the task. In
particular, we experiment with substituting the inflection generation component
with an LSTM sequence-to-sequence model and an LSTM pointer-generator network.
Our pointer-generator system obtains the best score of all seven submitted
systems on average over all languages, and outperforms the official baseline,
which was best overall, on Bulgarian and Kannada.
Related papers
- Heterogeneous Swarms: Jointly Optimizing Model Roles and Weights for Multi-LLM Systems [102.36545569092777]
We propose Heterogeneous Swarms, an algorithm to design multi-LLM systems by jointly optimizing model roles and weights.
Experiments demonstrate that Heterogeneous Swarms outperforms 15 role- and/or weight-based baselines by 18.5% on average across 12 tasks.
arXiv Detail & Related papers (2025-02-06T21:27:11Z) - RigLSTM: Recurrent Independent Grid LSTM for Generalizable Sequence
Learning [75.61681328968714]
We propose recurrent independent Grid LSTM (RigLSTM) to exploit the underlying modular structure of the target task.
Our model adopts cell selection, input feature selection, hidden state selection, and soft state updating to achieve a better generalization ability.
arXiv Detail & Related papers (2023-11-03T07:40:06Z) - Multi-Scale Prototypical Transformer for Whole Slide Image
Classification [12.584411225450989]
Whole slide image (WSI) classification is an essential task in computational pathology.
We propose a novel multi-scale prototypical Transformer (MSPT) for WSI classification.
arXiv Detail & Related papers (2023-07-05T14:10:29Z) - Coupling Global Context and Local Contents for Weakly-Supervised
Semantic Segmentation [54.419401869108846]
We propose a single-stage WeaklySupervised Semantic (WSSS) model with only the image-level class label supervisions.
A flexible context aggregation module is proposed to capture the global object context in different granular spaces.
A semantically consistent feature fusion module is proposed in a bottom-up parameter-learnable fashion to aggregate the fine-grained local contents.
arXiv Detail & Related papers (2023-04-18T15:29:23Z) - Revisiting Markovian Generative Architectures for Efficient
Task-Oriented Dialog Systems [22.249113574918034]
We propose to revisit Markovian Generative Architectures (MGA), which have been used in previous LSTM-based TOD systems.
Experiments on MultiWOZ2.1 show the efficiency advantages of the proposed Markovian PLM-based systems over their non-Markovian counterparts.
arXiv Detail & Related papers (2022-04-13T15:21:34Z) - Score-based Generative Modeling in Latent Space [93.8985523558869]
Score-based generative models (SGMs) have recently demonstrated impressive results in terms of both sample quality and distribution coverage.
Here, we propose the Latent Score-based Generative Model (LSGM), a novel approach that trains SGMs in a latent space.
Moving from data to latent space allows us to train more expressive generative models, apply SGMs to non-continuous data, and learn smoother SGMs in a smaller space.
arXiv Detail & Related papers (2021-06-10T17:26:35Z) - The LMU Munich System for the WMT 2020 Unsupervised Machine Translation
Shared Task [125.06737861979299]
This paper describes the submission of LMU Munich to the WMT 2020 unsupervised shared task, in two language directions.
Our core unsupervised neural machine translation (UNMT) system follows the strategy of Chronopoulou et al.
We ensemble our best-performing systems and reach a BLEU score of 32.4 on German->Upper Sorbian and 35.2 on Upper Sorbian->German.
arXiv Detail & Related papers (2020-10-25T19:04:03Z) - Revisiting LSTM Networks for Semi-Supervised Text Classification via
Mixed Objective Function [106.69643619725652]
We develop a training strategy that allows even a simple BiLSTM model, when trained with cross-entropy loss, to achieve competitive results.
We report state-of-the-art results for text classification task on several benchmark datasets.
arXiv Detail & Related papers (2020-09-08T21:55:22Z) - The NYU-CUBoulder Systems for SIGMORPHON 2020 Task 0 and Task 2 [25.234256237085336]
We describe the NYU-CUBoulder systems for the SIGMORPHON 2020 Task 0 on typologically diverse morphological inflection and Task 2 on unsupervised morphological paradigm completion.
The former consists of generating morphological inflections from a lemma and a set of morphosyntactic features describing the target form.
The latter requires generating entire paradigms for a set of given lemmas from raw text alone.
arXiv Detail & Related papers (2020-06-21T15:41:58Z) - Deep Autoencoding Topic Model with Scalable Hybrid Bayesian Inference [55.35176938713946]
We develop deep autoencoding topic model (DATM) that uses a hierarchy of gamma distributions to construct its multi-stochastic-layer generative network.
We propose a Weibull upward-downward variational encoder that deterministically propagates information upward via a deep neural network, followed by a downward generative model.
The efficacy and scalability of our models are demonstrated on both unsupervised and supervised learning tasks on big corpora.
arXiv Detail & Related papers (2020-06-15T22:22:56Z) - The SIGMORPHON 2020 Shared Task on Unsupervised Morphological Paradigm
Completion [28.728844366333185]
In this paper, we describe the findings of the SIGMORPHON 2020 shared task on unsupervised morphological paradigm completion.
Participants were asked to submit systems which take raw text and a list of lemmas as input, and output all inflected forms.
We present an analysis here, so that this shared task will ground further research on the topic.
arXiv Detail & Related papers (2020-05-28T03:09:58Z) - Yseop at SemEval-2020 Task 5: Cascaded BERT Language Model for
Counterfactual Statement Analysis [0.0]
We use a BERT base model for the classification task and build a hybrid BERT Multi-Layer Perceptron system to handle the sequence identification task.
Our experiments show that while introducing syntactic and semantic features does little in improving the system in the classification task, using these types of features as cascaded linear inputs to fine-tune the sequence-delimiting ability of the model ensures it outperforms other similar-purpose complex systems like BiLSTM-CRF in the second task.
arXiv Detail & Related papers (2020-05-18T08:19:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.