Semi-Supervised Neural System for Tagging, Parsing and Lematization
- URL: http://arxiv.org/abs/2004.12450v1
- Date: Sun, 26 Apr 2020 18:29:31 GMT
- Title: Semi-Supervised Neural System for Tagging, Parsing and Lematization
- Authors: Piotr Rybak, Alina Wr\'oblewska
- Abstract summary: This paper describes the ICS PAS system which took part in CoNLL 2018 shared task on Multilingual Parsing from Raw Text to Universal Dependencies.
The system consists of jointly trained tagger, lemmatizer, and dependency which are based on features extracted by a biLSTM network.
- Score: 1.6752182911522522
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper describes the ICS PAS system which took part in CoNLL 2018 shared
task on Multilingual Parsing from Raw Text to Universal Dependencies. The
system consists of jointly trained tagger, lemmatizer, and dependency parser
which are based on features extracted by a biLSTM network. The system uses both
fully connected and dilated convolutional neural architectures. The novelty of
our approach is the use of an additional loss function, which reduces the
number of cycles in the predicted dependency graphs, and the use of
self-training to increase the system performance. The proposed system, i.e. ICS
PAS (Warszawa), ranked 3th/4th in the official evaluation obtaining the
following overall results: 73.02 (LAS), 60.25 (MLAS) and 64.44 (BLEX).
Related papers
- MLCA-AVSR: Multi-Layer Cross Attention Fusion based Audio-Visual Speech Recognition [62.89464258519723]
We propose a multi-layer cross-attention fusion based AVSR approach that promotes representation of each modality by fusing them at different levels of audio/visual encoders.
Our proposed approach surpasses the first-place system, establishing a new SOTA cpCER of 29.13% on this dataset.
arXiv Detail & Related papers (2024-01-07T08:59:32Z) - Heterogenous Memory Augmented Neural Networks [84.29338268789684]
We introduce a novel heterogeneous memory augmentation approach for neural networks.
By introducing learnable memory tokens with attention mechanism, we can effectively boost performance without huge computational overhead.
We show our approach on various image and graph-based tasks under both in-distribution (ID) and out-of-distribution (OOD) conditions.
arXiv Detail & Related papers (2023-10-17T01:05:28Z) - A Performance-Consistent and Computation-Efficient CNN System for
High-Quality Automated Brain Tumor Segmentation [1.2183405753834562]
The research on developing CNN-based fully-automated Brain-Tumor-Segmentation systems has been progressed rapidly.
For the systems to be applicable in practice, a good processing quality and reliability are the must.
The CNN in the proposed system has a unique structure with 2 distinguished characters.
arXiv Detail & Related papers (2022-05-02T22:10:36Z) - Deliberation Model for On-Device Spoken Language Understanding [69.5587671262691]
We propose a novel deliberation-based approach to end-to-end (E2E) spoken language understanding (SLU)
We show that our approach can significantly reduce the degradation when moving from natural speech to synthetic speech training.
arXiv Detail & Related papers (2022-04-04T23:48:01Z) - Investigations on Speech Recognition Systems for Low-Resource Dialectal
Arabic-English Code-Switching Speech [32.426525641734344]
We present our work on code-switched Egyptian Arabic-English automatic speech recognition (ASR)
We build our ASR systems using DNN-based hybrid and Transformer-based end-to-end models.
We show that recognition can be improved by combining the outputs of both systems.
arXiv Detail & Related papers (2021-08-29T17:23:30Z) - Low Resource German ASR with Untranscribed Data Spoken by Non-native
Children -- INTERSPEECH 2021 Shared Task SPAPL System [19.435571932141364]
This paper describes the SPAPL system for the INTERSPEECH 2021 Challenge: Shared Task on Automatic Speech Recognition for Non-Native Children's Speech in German.
5 hours of transcribed data and 60 hours of untranscribed data are provided to develop a German ASR system for children.
For the training of the transcribed data, we propose a non-speech state discriminative loss (NSDL) to mitigate the influence of long-duration non-speech segments within speech utterances.
Our system achieves a word error rate (WER) of 39.68% on the evaluation data,
arXiv Detail & Related papers (2021-06-18T07:36:26Z) - SpanNer: Named Entity Re-/Recognition as Span Prediction [62.66148736099347]
span prediction model is used for named entity recognition.
We experimentally implement 154 systems on 11 datasets, covering three languages.
Our model has been deployed into the ExplainaBoard platform.
arXiv Detail & Related papers (2021-06-01T17:11:42Z) - Coreference Resolution System for Indonesian Text with Mention Pair
Method and Singleton Exclusion using Convolutional Neural Network [0.0]
We propose a new coreference resolution system for Indonesian text with mention pair method.
In addition to lexical and syntactic features, in order to learn the representation of the mentions words and context, we use word embeddings and feed them to CNN.
Our proposed system outperforms the state-of-the-art system.
arXiv Detail & Related papers (2020-09-11T22:21:19Z) - Neural Architecture Search For LF-MMI Trained Time Delay Neural Networks [61.76338096980383]
A range of neural architecture search (NAS) techniques are used to automatically learn two types of hyper- parameters of state-of-the-art factored time delay neural networks (TDNNs)
These include the DARTS method integrating architecture selection with lattice-free MMI (LF-MMI) TDNN training.
Experiments conducted on a 300-hour Switchboard corpus suggest the auto-configured systems consistently outperform the baseline LF-MMI TDNN systems.
arXiv Detail & Related papers (2020-07-17T08:32:11Z) - Jointly Trained Transformers models for Spoken Language Translation [2.3886615435250302]
This work trains SLT systems with ASR objective as an auxiliary loss and both the networks are connected through neural hidden representations.
This architecture has improved from BLEU from 36.8 to 44.5.
All the experiments are reported on English-Portuguese speech translation task using How2 corpus.
arXiv Detail & Related papers (2020-04-25T11:28:39Z) - ResNeSt: Split-Attention Networks [86.25490825631763]
We present a modularized architecture, which applies the channel-wise attention on different network branches to leverage their success in capturing cross-feature interactions and learning diverse representations.
Our model, named ResNeSt, outperforms EfficientNet in accuracy and latency trade-off on image classification.
arXiv Detail & Related papers (2020-04-19T20:40:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.