Style Variation as a Vantage Point for Code-Switching
- URL: http://arxiv.org/abs/2005.00458v1
- Date: Fri, 1 May 2020 15:53:16 GMT
- Title: Style Variation as a Vantage Point for Code-Switching
- Authors: Khyathi Raghavi Chandu, Alan W Black
- Abstract summary: Code-Switching (CS) is a common phenomenon observed in several bilingual and multilingual communities.
We present a novel vantage point of CS to be style variations between both the participating languages.
We propose a two-stage generative adversarial training approach where the first stage generates competitive negative examples for CS and the second stage generates more realistic CS sentences.
- Score: 54.34370423151014
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Code-Switching (CS) is a common phenomenon observed in several bilingual and
multilingual communities, thereby attaining prevalence in digital and social
media platforms. This increasing prominence demands the need to model CS
languages for critical downstream tasks. A major problem in this domain is the
dearth of annotated data and a substantial corpora to train large scale neural
models. Generating vast amounts of quality text assists several down stream
tasks that heavily rely on language modeling such as speech recognition,
text-to-speech synthesis etc,. We present a novel vantage point of CS to be
style variations between both the participating languages. Our approach does
not need any external annotations such as lexical language ids. It mainly
relies on easily obtainable monolingual corpora without any parallel alignment
and a limited set of naturally CS sentences. We propose a two-stage generative
adversarial training approach where the first stage generates competitive
negative examples for CS and the second stage generates more realistic CS
sentences. We present our experiments on the following pairs of languages:
Spanish-English, Mandarin-English, Hindi-English and Arabic-French. We show
that the trends in metrics for generated CS move closer to real CS data in each
of the above language pairs through the dual stage training process. We believe
this viewpoint of CS as style variations opens new perspectives for modeling
various tasks in CS text.
Related papers
- ConCSE: Unified Contrastive Learning and Augmentation for Code-Switched Embeddings [4.68732641979009]
This paper examines the Code-Switching (CS) phenomenon where two languages intertwine within a single utterance.
We highlight that the current Equivalence Constraint (EC) theory for CS in other languages may only partially capture English-Korean CS complexities.
We introduce a novel Koglish dataset tailored for English-Korean CS scenarios to mitigate such challenges.
arXiv Detail & Related papers (2024-08-28T11:27:21Z) - Bridge to Non-Barrier Communication: Gloss-Prompted Fine-grained Cued Speech Gesture Generation with Diffusion Model [11.160802635050866]
Cued Speech (CS) is an advanced visual phonetic encoding system that integrates lip reading with hand codings.
Existing CS generation methods are fragile and prone to poor performance due to template-based statistical models.
We propose a novel Gloss-prompted Diffusion-based CS Gesture generation framework (called GlossDiff)
arXiv Detail & Related papers (2024-04-30T05:54:40Z) - Language-agnostic Code-Switching in Sequence-To-Sequence Speech
Recognition [62.997667081978825]
Code-Switching (CS) is referred to the phenomenon of alternately using words and phrases from different languages.
We propose a simple yet effective data augmentation in which audio and corresponding labels of different source languages are transcribed.
We show that this augmentation can even improve the model's performance on inter-sentential language switches not seen during training by 5,03% WER.
arXiv Detail & Related papers (2022-10-17T12:15:57Z) - The Who in Code-Switching: A Case Study for Predicting Egyptian
Arabic-English Code-Switching Levels based on Character Profiles [20.746558640332953]
Code-switching (CS) is a linguistic phenomenon exhibited by multilingual individuals, where they tend to alternate between languages within one single conversation.
We use machine learning (ML) to predict users' CS levels based on their profiles.
Our results show that the CS behaviour is affected by the relation between speakers, travel experiences as well as Neuroticism and Extraversion personality traits.
arXiv Detail & Related papers (2022-07-31T13:47:35Z) - End-to-End Speech Translation for Code Switched Speech [13.97982457879585]
Code switching (CS) refers to the phenomenon of interchangeably using words and phrases from different languages.
We focus on CS in the context of English/Spanish conversations for the task of speech translation (ST), generating and evaluating both transcript and translation.
We show that our ST architectures, and especially our bidirectional end-to-end architecture, perform well on CS speech, even when no CS training data is used.
arXiv Detail & Related papers (2022-04-11T13:25:30Z) - Exploring Teacher-Student Learning Approach for Multi-lingual
Speech-to-Intent Classification [73.5497360800395]
We develop an end-to-end system that supports multiple languages.
We exploit knowledge from a pre-trained multi-lingual natural language processing model.
arXiv Detail & Related papers (2021-09-28T04:43:11Z) - Discovering Representation Sprachbund For Multilingual Pre-Training [139.05668687865688]
We generate language representation from multilingual pre-trained models and conduct linguistic analysis.
We cluster all the target languages into multiple groups and name each group as a representation sprachbund.
Experiments are conducted on cross-lingual benchmarks and significant improvements are achieved compared to strong baselines.
arXiv Detail & Related papers (2021-09-01T09:32:06Z) - Towards Zero-shot Language Modeling [90.80124496312274]
We construct a neural model that is inductively biased towards learning human languages.
We infer this distribution from a sample of typologically diverse training languages.
We harness additional language-specific side information as distant supervision for held-out languages.
arXiv Detail & Related papers (2021-08-06T23:49:18Z) - KARI: KAnari/QCRI's End-to-End systems for the INTERSPEECH 2021 Indian
Languages Code-Switching Challenge [7.711092265101041]
We present the Kanari/QCRI system and the modeling strategies used to participate in the Interspeech 2021 Code-switching (CS) challenge for low-resource Indian languages.
The subtask involved developing a speech recognition system for two CS datasets: Hindi-English and Bengali-English.
To tackle the CS challenges, we use transfer learning for incorporating the publicly available monolingual Hindi, Bengali, and English speech data.
arXiv Detail & Related papers (2021-06-10T16:12:51Z) - Vokenization: Improving Language Understanding with Contextualized,
Visual-Grounded Supervision [110.66085917826648]
We develop a technique that extrapolates multimodal alignments to language-only data by contextually mapping language tokens to their related images.
"vokenization" is trained on relatively small image captioning datasets and we then apply it to generate vokens for large language corpora.
Trained with these contextually generated vokens, our visually-supervised language models show consistent improvements over self-supervised alternatives on multiple pure-language tasks.
arXiv Detail & Related papers (2020-10-14T02:11:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.