Applying Automated Machine Translation to Educational Video Courses
- URL: http://arxiv.org/abs/2301.03141v2
- Date: Tue, 19 Sep 2023 01:16:00 GMT
- Title: Applying Automated Machine Translation to Educational Video Courses
- Authors: Linden Wang
- Abstract summary: We studied the capability of automated machine translation in the online video education space.
We applied text-to-speech synthesis and audio/video synchronization to build engaging videos in target languages.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We studied the capability of automated machine translation in the online
video education space by automatically translating Khan Academy videos with
state-of-the-art translation models and applying text-to-speech synthesis and
audio/video synchronization to build engaging videos in target languages. We
also analyzed and established two reliable translation confidence estimators
based on round-trip translations in order to efficiently manage translation
quality and reduce human translation effort. Finally, we developed a deployable
system to deliver translated videos to end users and collect user corrections
for iterative improvement.
Related papers
- Bilingual Corpus Mining and Multistage Fine-Tuning for Improving Machine
Translation of Lecture Transcripts [50.00305136008848]
We propose a framework for parallel corpus mining, which provides a quick and effective way to mine a parallel corpus from publicly available lectures on Coursera.
For both English--Japanese and English--Chinese lecture translations, we extracted parallel corpora of approximately 50,000 lines and created development and test sets.
This study also suggests guidelines for gathering and cleaning corpora, mining parallel sentences, cleaning noise in the mined data, and creating high-quality evaluation splits.
arXiv Detail & Related papers (2023-11-07T03:50:25Z) - Synslator: An Interactive Machine Translation Tool with Online Learning [19.6690111454278]
Interactive machine translation (IMT) has emerged as a progression of the computer-aided translation paradigm.
This paper introduces Synslator, a user-friendly computer-aided translation (CAT) tool that not only supports IMT, but is adept at online learning with real-time translation memories.
arXiv Detail & Related papers (2023-10-08T06:05:55Z) - Video-Teller: Enhancing Cross-Modal Generation with Fusion and
Decoupling [79.49128866877922]
Video-Teller is a video-language foundation model that leverages multi-modal fusion and fine-grained modality alignment.
Video-Teller boosts the training efficiency by utilizing frozen pretrained vision and language modules.
It capitalizes on the robust linguistic capabilities of large language models, enabling the generation of both concise and elaborate video descriptions.
arXiv Detail & Related papers (2023-10-08T03:35:27Z) - Incorporating Human Translator Style into English-Turkish Literary
Machine Translation [0.26168876987285306]
We develop machine translation models that take into account the stylistic features of translators.
We show that the human translator style can be highly recreated in the target machine translations.
arXiv Detail & Related papers (2023-07-21T09:39:50Z) - Improving Isochronous Machine Translation with Target Factors and
Auxiliary Counters [71.02335065794384]
We introduce target factors in a transformer model to predict durations jointly with target language phoneme sequences.
We show that our model improves translation quality and isochrony compared to previous work.
arXiv Detail & Related papers (2023-05-22T16:36:04Z) - Decomposed Prompting for Machine Translation Between Related Languages
using Large Language Models [55.35106713257871]
We introduce DecoMT, a novel approach of few-shot prompting that decomposes the translation process into a sequence of word chunk translations.
We show that DecoMT outperforms the strong few-shot prompting BLOOM model with an average improvement of 8 chrF++ scores across the examined languages.
arXiv Detail & Related papers (2023-05-22T14:52:47Z) - The Best of Both Worlds: Combining Human and Machine Translations for
Multilingual Semantic Parsing with Active Learning [50.320178219081484]
We propose an active learning approach that exploits the strengths of both human and machine translations.
An ideal utterance selection can significantly reduce the error and bias in the translated data.
arXiv Detail & Related papers (2023-05-22T05:57:47Z) - BSTC: A Large-Scale Chinese-English Speech Translation Dataset [26.633433687767553]
BSTC (Baidu Speech Translation Corpus) is a large-scale Chinese-English speech translation dataset.
This dataset is constructed based on a collection of licensed videos of talks or lectures, including about 68 hours of Mandarin data.
We have asked three experienced interpreters to simultaneously interpret the testing talks in a mock conference setting.
arXiv Detail & Related papers (2021-04-08T07:38:51Z) - Designing the Business Conversation Corpus [20.491255702901288]
We aim to boost the machine translation quality of conversational texts by introducing a newly constructed Japanese-English business conversation parallel corpus.
A detailed analysis of the corpus is provided along with challenging examples for automatic translation.
We also experiment with adding the corpus in a machine translation training scenario and show how the resulting system benefits from its use.
arXiv Detail & Related papers (2020-08-05T05:19:44Z) - Keyframe Segmentation and Positional Encoding for Video-guided Machine
Translation Challenge 2020 [28.38178018722211]
We presented our video-guided machine translation system in approaching the Video-guided Machine Translation Challenge 2020.
In the evaluation phase, our system scored 36.60 corpus-level BLEU-4 and achieved the 1st place on the Video-guided Machine Translation Challenge 2020.
arXiv Detail & Related papers (2020-06-23T07:15:11Z) - Unsupervised Multimodal Video-to-Video Translation via Self-Supervised
Learning [92.17835753226333]
We propose a novel unsupervised video-to-video translation model.
Our model decomposes the style and the content using the specialized UV-decoder structure.
Our model can produce photo-realistic videos in a multimodal way.
arXiv Detail & Related papers (2020-04-14T13:44:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.