An End-to-end Chinese Text Normalization Model based on Rule-guided
Flat-Lattice Transformer
- URL: http://arxiv.org/abs/2203.16954v1
- Date: Thu, 31 Mar 2022 11:19:53 GMT
- Title: An End-to-end Chinese Text Normalization Model based on Rule-guided
Flat-Lattice Transformer
- Authors: Wenlin Dai, Changhe Song, Xiang Li, Zhiyong Wu, Huashan Pan, Xiulin
Li, Helen Meng
- Abstract summary: We propose an end-to-end Chinese text normalization model, which accepts Chinese characters as direct input.
We also release a first publicly accessible largescale dataset for Chinese text normalization.
- Score: 37.0774363352316
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Text normalization, defined as a procedure transforming non standard words to
spoken-form words, is crucial to the intelligibility of synthesized speech in
text-to-speech system. Rule-based methods without considering context can not
eliminate ambiguation, whereas sequence-to-sequence neural network based
methods suffer from the unexpected and uninterpretable errors problem. Recently
proposed hybrid system treats rule-based model and neural model as two cascaded
sub-modules, where limited interaction capability makes neural network model
cannot fully utilize expert knowledge contained in the rules. Inspired by
Flat-LAttice Transformer (FLAT), we propose an end-to-end Chinese text
normalization model, which accepts Chinese characters as direct input and
integrates expert knowledge contained in rules into the neural network, both
contribute to the superior performance of proposed model for the text
normalization task. We also release a first publicly accessible largescale
dataset for Chinese text normalization. Our proposed model has achieved
excellent results on this dataset.
Related papers
- A Transformer-based Approach for Arabic Offline Handwritten Text
Recognition [0.0]
We introduce two alternative architectures for recognizing offline Arabic handwritten text.
Our approach can model language dependencies and relies only on the attention mechanism, thereby making it more parallelizable and less complex.
Our evaluation on the Arabic KHATT dataset demonstrates that our proposed method outperforms the current state-of-the-art approaches.
arXiv Detail & Related papers (2023-07-27T17:51:52Z) - The Whole Truth and Nothing But the Truth: Faithful and Controllable
Dialogue Response Generation with Dataflow Transduction and Constrained
Decoding [65.34601470417967]
We describe a hybrid architecture for dialogue response generation that combines the strengths of neural language modeling and rule-based generation.
Our experiments show that this system outperforms both rule-based and learned approaches in human evaluations of fluency, relevance, and truthfulness.
arXiv Detail & Related papers (2022-09-16T09:00:49Z) - Shallow Fusion of Weighted Finite-State Transducer and Language Model
for Text Normalization [13.929356163132558]
We propose a new hybrid approach that combines the benefits of rule-based and neural systems.
First, a non-deterministic WFST outputs all normalization candidates, and then a neural language model picks the best one.
It achieves comparable or better results than existing state-of-the-art TN models.
arXiv Detail & Related papers (2022-03-29T21:34:35Z) - Factorized Neural Transducer for Efficient Language Model Adaptation [51.81097243306204]
We propose a novel model, factorized neural Transducer, by factorizing the blank and vocabulary prediction.
It is expected that this factorization can transfer the improvement of the standalone language model to the Transducer for speech recognition.
We demonstrate that the proposed factorized neural Transducer yields 15% to 20% WER improvements when out-of-domain text data is used for language model adaptation.
arXiv Detail & Related papers (2021-09-27T15:04:00Z) - Sequence-to-Sequence Learning with Latent Neural Grammars [12.624691611049341]
Sequence-to-sequence learning with neural networks has become the de facto standard for sequence prediction tasks.
While flexible and performant, these models often require large datasets for training and can fail spectacularly on benchmarks designed to test for compositional generalization.
This work explores an alternative, hierarchical approach to sequence-to-sequence learning with quasi-synchronous grammars.
arXiv Detail & Related papers (2021-09-02T17:58:08Z) - Structured Reordering for Modeling Latent Alignments in Sequence
Transduction [86.94309120789396]
We present an efficient dynamic programming algorithm performing exact marginal inference of separable permutations.
The resulting seq2seq model exhibits better systematic generalization than standard models on synthetic problems and NLP tasks.
arXiv Detail & Related papers (2021-06-06T21:53:54Z) - Neural Inverse Text Normalization [11.240669509034298]
We propose an efficient and robust neural solution for inverse text normalization.
We show that this can be easily extended to other languages without the need for a linguistic expert to manually curate them.
A transformer based model infused with pretraining consistently achieves a lower WER across several datasets.
arXiv Detail & Related papers (2021-02-12T07:53:53Z) - GTAE: Graph-Transformer based Auto-Encoders for Linguistic-Constrained
Text Style Transfer [119.70961704127157]
Non-parallel text style transfer has attracted increasing research interests in recent years.
Current approaches still lack the ability to preserve the content and even logic of original sentences.
We propose a method called Graph Transformer based Auto-GTAE, which models a sentence as a linguistic graph and performs feature extraction and style transfer at the graph level.
arXiv Detail & Related papers (2021-02-01T11:08:45Z) - DiscreTalk: Text-to-Speech as a Machine Translation Problem [52.33785857500754]
This paper proposes a new end-to-end text-to-speech (E2E-TTS) model based on neural machine translation (NMT)
The proposed model consists of two components; a non-autoregressive vector quantized variational autoencoder (VQ-VAE) model and an autoregressive Transformer-NMT model.
arXiv Detail & Related papers (2020-05-12T02:45:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.