MRN: Multiplexed Routing Network for Incremental Multilingual Text
Recognition
- URL: http://arxiv.org/abs/2305.14758v3
- Date: Sun, 30 Jul 2023 07:40:29 GMT
- Title: MRN: Multiplexed Routing Network for Incremental Multilingual Text
Recognition
- Authors: Tianlun Zheng, Zhineng Chen, BingChen Huang, Wei Zhang and Yu-Gang
Jiang
- Abstract summary: Multiplexed routing network (MRN) trains a recognizer for each language that is currently seen.
MRN effectively reduces the reliance on older data and better fights against catastrophic forgetting.
It outperforms existing general-purpose IL methods by large margins.
- Score: 56.408324994409405
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multilingual text recognition (MLTR) systems typically focus on a fixed set
of languages, which makes it difficult to handle newly added languages or adapt
to ever-changing data distribution. In this paper, we propose the Incremental
MLTR (IMLTR) task in the context of incremental learning (IL), where different
languages are introduced in batches. IMLTR is particularly challenging due to
rehearsal-imbalance, which refers to the uneven distribution of sample
characters in the rehearsal set, used to retain a small amount of old data as
past memories. To address this issue, we propose a Multiplexed Routing Network
(MRN). MRN trains a recognizer for each language that is currently seen.
Subsequently, a language domain predictor is learned based on the rehearsal set
to weigh the recognizers. Since the recognizers are derived from the original
data, MRN effectively reduces the reliance on older data and better fights
against catastrophic forgetting, the core issue in IL. We extensively evaluate
MRN on MLT17 and MLT19 datasets. It outperforms existing general-purpose IL
methods by large margins, with average accuracy improvements ranging from 10.3%
to 35.8% under different settings. Code is available at
https://github.com/simplify23/MRN.
Related papers
- Integrating Multi-scale Contextualized Information for Byte-based Neural Machine Translation [14.826948179996695]
Subword tokenization is a common method for vocabulary building in Neural Machine Translation (NMT) models.
We propose the Multi-Scale Contextualization (MSC) method, which learns contextualized information of varying scales across different hidden state dimensions.
Experiments show that MSC significantly outperforms subword-based and other byte-based methods in both multilingual and out-of-domain scenarios.
arXiv Detail & Related papers (2024-05-29T17:19:04Z) - Chain-of-Dictionary Prompting Elicits Translation in Large Language Models [100.47154959254937]
Large language models (LLMs) have shown surprisingly good performance in multilingual neural machine translation (MNMT)
We present a novel method, CoD, which augments LLMs with prior knowledge with the chains of multilingual dictionaries for a subset of input words to elicit translation abilities.
arXiv Detail & Related papers (2023-05-11T05:19:47Z) - Boosting Zero-shot Cross-lingual Retrieval by Training on Artificially
Code-Switched Data [26.38449396649045]
We show that the effectiveness of zero-shot rankers diminishes when queries and documents are present in different languages.
Motivated by this, we propose to train ranking models on artificially code-switched data instead.
arXiv Detail & Related papers (2023-05-09T09:32:19Z) - Memory Augmented Lookup Dictionary based Language Modeling for Automatic
Speech Recognition [20.926163659469587]
We propose a new memory augmented lookup dictionary based Transformer architecture for LM.
The newly introduced lookup dictionary incorporates rich contextual information in training set, which is vital to correctly predict long-tail tokens.
Our proposed method is proved to outperform the baseline Transformer LM by a great margin on both word/character error rate and tail tokens error rate.
arXiv Detail & Related papers (2022-12-30T22:26:57Z) - Few-Shot Cross-lingual Transfer for Coarse-grained De-identification of
Code-Mixed Clinical Texts [56.72488923420374]
Pre-trained language models (LMs) have shown great potential for cross-lingual transfer in low-resource settings.
We show the few-shot cross-lingual transfer property of LMs for named recognition (NER) and apply it to solve a low-resource and real-world challenge of code-mixed (Spanish-Catalan) clinical notes de-identification in the stroke.
arXiv Detail & Related papers (2022-04-10T21:46:52Z) - Parameter-Efficient Neural Reranking for Cross-Lingual and Multilingual
Retrieval [66.69799641522133]
State-of-the-art neural (re)rankers are notoriously data hungry.
Current approaches typically transfer rankers trained on English data to other languages and cross-lingual setups by means of multilingual encoders.
We show that two parameter-efficient approaches to cross-lingual transfer, namely Sparse Fine-Tuning Masks (SFTMs) and Adapters, allow for a more lightweight and more effective zero-shot transfer.
arXiv Detail & Related papers (2022-04-05T15:44:27Z) - Low-Resource Spoken Language Identification Using Self-Attentive Pooling
and Deep 1D Time-Channel Separable Convolutions [0.0]
We show that a convolutional neural network with a Self-Attentive Pooling layer shows promising results in low-resource setting for the language identification task.
We also substantiate the hypothesis that whenever the dataset is diverse enough so that the other classification factors, like gender, age etc. are well-averaged, the confusion matrix for LID system bears the language similarity measure.
arXiv Detail & Related papers (2021-05-31T18:35:27Z) - Explicit Alignment Objectives for Multilingual Bidirectional Encoders [111.65322283420805]
We present a new method for learning multilingual encoders, AMBER (Aligned Multilingual Bi-directional EncodeR)
AMBER is trained on additional parallel data using two explicit alignment objectives that align the multilingual representations at different granularities.
Experimental results show that AMBER obtains gains of up to 1.1 average F1 score on sequence tagging and up to 27.3 average accuracy on retrieval over the XLMR-large model.
arXiv Detail & Related papers (2020-10-15T18:34:13Z) - X-FACTR: Multilingual Factual Knowledge Retrieval from Pretrained
Language Models [103.75890012041366]
Language models (LMs) have proven surprisingly successful at capturing factual knowledge.
However, studies on LMs' factual representation ability have almost invariably been performed on English.
We create a benchmark of cloze-style probes for 23 typologically diverse languages.
arXiv Detail & Related papers (2020-10-13T05:29:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.