Improving speech translation by fusing speech and text
- URL: http://arxiv.org/abs/2305.14042v1
- Date: Tue, 23 May 2023 13:13:48 GMT
- Title: Improving speech translation by fusing speech and text
- Authors: Wenbiao Yin, Zhicheng Liu, Chengqi Zhao, Tao Wang, Jian Tong, Rong Ye
- Abstract summary: We harness the complementary strengths of speech and text, which are disparate modalities.
We propose textbfFuse-textbfSpeech-textbfText (textbfFST), a cross-modal model which supports three distinct input modalities for translation.
- Score: 24.31233927318388
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In speech translation, leveraging multimodal data to improve model
performance and address limitations of individual modalities has shown
significant effectiveness. In this paper, we harness the complementary
strengths of speech and text, which are disparate modalities. We observe three
levels of modality gap between them, denoted by Modal input representation,
Modal semantic, and Modal hidden states. To tackle these gaps, we propose
\textbf{F}use-\textbf{S}peech-\textbf{T}ext (\textbf{FST}), a cross-modal model
which supports three distinct input modalities for translation: speech, text,
and fused speech-text. We leverage multiple techniques for cross-modal
alignment and conduct a comprehensive analysis to assess its impact on speech
translation, machine translation, and fused speech-text translation. We
evaluate FST on MuST-C, GigaST, and newstest benchmark. Experiments show that
the proposed FST achieves an average 34.0 BLEU on MuST-C
En$\rightarrow$De/Es/Fr (vs SOTA +1.1 BLEU). Further experiments demonstrate
that FST does not degrade on MT task, as observed in prior works. Instead, it
yields an average improvement of 3.2 BLEU over the pre-trained MT model.
Related papers
- FASST: Fast LLM-based Simultaneous Speech Translation [9.65638081954595]
Simultaneous speech translation (SST) takes streaming speech input and generates text translation on the fly.
We propose FASST, a fast large language model based method for streaming speech translation.
Experiment results show that FASST achieves the best quality-latency trade-off.
arXiv Detail & Related papers (2024-08-18T10:12:39Z) - Cross-Modal Multi-Tasking for Speech-to-Text Translation via Hard
Parameter Sharing [72.56219471145232]
We propose a ST/MT multi-tasking framework with hard parameter sharing.
Our method reduces the speech-text modality gap via a pre-processing stage.
We show that our framework improves attentional encoder-decoder, Connectionist Temporal Classification (CTC), transducer, and joint CTC/attention models by an average of +0.5 BLEU.
arXiv Detail & Related papers (2023-09-27T17:48:14Z) - Beyond Triplet: Leveraging the Most Data for Multimodal Machine
Translation [53.342921374639346]
Multimodal machine translation aims to improve translation quality by incorporating information from other modalities, such as vision.
Previous MMT systems mainly focus on better access and use of visual information and tend to validate their methods on image-related datasets.
This paper establishes new methods and new datasets for MMT.
arXiv Detail & Related papers (2022-12-20T15:02:38Z) - M3ST: Mix at Three Levels for Speech Translation [66.71994367650461]
We propose Mix at three levels for Speech Translation (M3ST) method to increase the diversity of the augmented training corpus.
In the first stage of fine-tuning, we mix the training corpus at three levels, including word level, sentence level and frame level, and fine-tune the entire model with mixed data.
Experiments on MuST-C speech translation benchmark and analysis show that M3ST outperforms current strong baselines and achieves state-of-the-art results on eight directions with an average BLEU of 29.9.
arXiv Detail & Related papers (2022-12-07T14:22:00Z) - Simple and Effective Unsupervised Speech Translation [68.25022245914363]
We study a simple and effective approach to build speech translation systems without labeled data.
We present an unsupervised domain adaptation technique for pre-trained speech models.
Experiments show that unsupervised speech-to-text translation outperforms the previous unsupervised state of the art.
arXiv Detail & Related papers (2022-10-18T22:26:13Z) - Revamping Multilingual Agreement Bidirectionally via Switched
Back-translation for Multilingual Neural Machine Translation [107.83158521848372]
multilingual agreement (MA) has shown its importance for multilingual neural machine translation (MNMT)
We present textbfBidirectional textbfMultilingual textbfAgreement via textbfSwitched textbfBack-textbftranslation (textbfBMA-SBT)
It is a novel and universal multilingual agreement framework for fine-tuning pre-trained MNMT models.
arXiv Detail & Related papers (2022-09-28T09:14:58Z) - STEMM: Self-learning with Speech-text Manifold Mixup for Speech
Translation [37.51435498386953]
We propose the Speech-TExt Manifold Mixup (STEMM) method to calibrate such discrepancy.
Experiments on MuST-C speech translation benchmark and further analysis show that our method effectively alleviates the cross-modal representation discrepancy.
arXiv Detail & Related papers (2022-03-20T01:49:53Z) - Learning Shared Semantic Space for Speech-to-Text Translation [32.12445734213848]
We propose to bridge the modality gap between text machine translation (MT) and end-to-end speech translation (ST)
By projecting audio and text features to a common semantic representation, Chimera unifies MT and ST tasks.
Specifically, Chimera obtains 26.3 BLEU on EN-DE, improving the SOTA by a +2.7 BLEU margin.
arXiv Detail & Related papers (2021-05-07T07:49:56Z) - End-to-end Speech Translation via Cross-modal Progressive Training [12.916100727707809]
Cross Speech-Text Network (XSTNet) is an end-to-end model for speech-to-text translation.
XSTNet takes both speech and text as input and outputs both transcription and translation text.
XSTNet achieves state-of-the-art results on all three language directions with an average BLEU of 27.8, outperforming the previous best method by 3.7 BLEU.
arXiv Detail & Related papers (2021-04-21T06:44:31Z) - Consecutive Decoding for Speech-to-text Translation [51.155661276936044]
COnSecutive Transcription and Translation (COSTT) is an integral approach for speech-to-text translation.
The key idea is to generate source transcript and target translation text with a single decoder.
Our method is verified on three mainstream datasets.
arXiv Detail & Related papers (2020-09-21T10:10:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.