Neural Inverse Text Normalization
- URL: http://arxiv.org/abs/2102.06380v1
- Date: Fri, 12 Feb 2021 07:53:53 GMT
- Title: Neural Inverse Text Normalization
- Authors: Monica Sunkara, Chaitanya Shivade, Sravan Bodapati, Katrin Kirchhoff
- Abstract summary: We propose an efficient and robust neural solution for inverse text normalization.
We show that this can be easily extended to other languages without the need for a linguistic expert to manually curate them.
A transformer based model infused with pretraining consistently achieves a lower WER across several datasets.
- Score: 11.240669509034298
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: While there have been several contributions exploring state of the art
techniques for text normalization, the problem of inverse text normalization
(ITN) remains relatively unexplored. The best known approaches leverage finite
state transducer (FST) based models which rely on manually curated rules and
are hence not scalable. We propose an efficient and robust neural solution for
ITN leveraging transformer based seq2seq models and FST-based text
normalization techniques for data preparation. We show that this can be easily
extended to other languages without the need for a linguistic expert to
manually curate them. We then present a hybrid framework for integrating Neural
ITN with an FST to overcome common recoverable errors in production
environments. Our empirical evaluations show that the proposed solution
minimizes incorrect perturbations (insertions, deletions and substitutions) to
ASR output and maintains high quality even on out of domain data. A transformer
based model infused with pretraining consistently achieves a lower WER across
several datasets and is able to outperform baselines on English, Spanish,
German and Italian datasets.
Related papers
- Optimizing Non-Autoregressive Transformers with Contrastive Learning [74.46714706658517]
Non-autoregressive Transformers (NATs) reduce the inference latency of Autoregressive Transformers (ATs) by predicting words all at once rather than in sequential order.
In this paper, we propose to ease the difficulty of modality learning via sampling from the model distribution instead of the data distribution.
arXiv Detail & Related papers (2023-05-23T04:20:13Z) - Categorizing Semantic Representations for Neural Machine Translation [53.88794787958174]
We introduce categorization to the source contextualized representations.
The main idea is to enhance generalization by reducing sparsity and overfitting.
Experiments on a dedicated MT dataset show that our method reduces compositional generalization error rates by 24% error reduction.
arXiv Detail & Related papers (2022-10-13T04:07:08Z) - Thutmose Tagger: Single-pass neural model for Inverse Text Normalization [76.87664008338317]
Inverse text normalization (ITN) is an essential post-processing step in automatic speech recognition.
We present a dataset preparation method based on the granular alignment of ITN examples.
One-to-one correspondence between tags and input words improves the interpretability of the model's predictions.
arXiv Detail & Related papers (2022-07-29T20:39:02Z) - Non-Parametric Domain Adaptation for End-to-End Speech Translation [72.37869362559212]
End-to-End Speech Translation (E2E-ST) has received increasing attention due to the potential of its less error propagation, lower latency, and fewer parameters.
We propose a novel non-parametric method that leverages domain-specific text translation corpus to achieve domain adaptation for the E2E-ST system.
arXiv Detail & Related papers (2022-05-23T11:41:02Z) - Learning to Generalize to More: Continuous Semantic Augmentation for
Neural Machine Translation [50.54059385277964]
We present a novel data augmentation paradigm termed Continuous Semantic Augmentation (CsaNMT)
CsaNMT augments each training instance with an adjacency region that could cover adequate variants of literal expression under the same meaning.
arXiv Detail & Related papers (2022-04-14T08:16:28Z) - Shallow Fusion of Weighted Finite-State Transducer and Language Model
for Text Normalization [13.929356163132558]
We propose a new hybrid approach that combines the benefits of rule-based and neural systems.
First, a non-deterministic WFST outputs all normalization candidates, and then a neural language model picks the best one.
It achieves comparable or better results than existing state-of-the-art TN models.
arXiv Detail & Related papers (2022-03-29T21:34:35Z) - Sequence-to-Sequence Lexical Normalization with Multilingual
Transformers [3.3302293148249125]
Current benchmark tasks for natural language processing contain text that is qualitatively different from the text used in informal day to day digital communication.
This discrepancy has led to severe performance degradation of state-of-the-art NLP models when fine-tuned on real-world data.
We propose a sentence-level sequence-to-sequence model based on mBART, which frames the problem as a machine translation problem.
arXiv Detail & Related papers (2021-10-06T15:53:20Z) - A Unified Transformer-based Framework for Duplex Text Normalization [33.90810154067128]
Text normalization (TN) and inverse text normalization (ITN) are essential preprocessing and postprocessing steps for text-to-speech synthesis and automatic speech recognition.
We propose a unified framework for building a single neural duplex system that can simultaneously handle TN and ITN.
Our systems achieve state-of-the-art results on the Google TN dataset for English and Russian.
arXiv Detail & Related papers (2021-08-23T01:55:03Z) - SDA: Improving Text Generation with Self Data Augmentation [88.24594090105899]
We propose to improve the standard maximum likelihood estimation (MLE) paradigm by incorporating a self-imitation-learning phase for automatic data augmentation.
Unlike most existing sentence-level augmentation strategies, our method is more general and could be easily adapted to any MLE-based training procedure.
arXiv Detail & Related papers (2021-01-02T01:15:57Z) - Sentence Boundary Augmentation For Neural Machine Translation Robustness [11.290581889247983]
We show that sentence boundary segmentation has the largest impact on quality, and we develop a simple data augmentation strategy to improve segmentation robustness.
We show that sentence boundary segmentation has the largest impact on quality, and we develop a simple data augmentation strategy to improve segmentation robustness.
arXiv Detail & Related papers (2020-10-21T16:44:48Z) - Variational Neural Machine Translation with Normalizing Flows [13.537869825364718]
Variational Neural Machine Translation (VNMT) is an attractive framework for modeling the generation of target translations.
We propose to apply the VNMT framework to the state-of-the-art Transformer and introduce a more flexible approximate posterior based on normalizing flows.
arXiv Detail & Related papers (2020-05-28T13:30:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.