From Unsupervised Machine Translation To Adversarial Text Generation
- URL: http://arxiv.org/abs/2011.05449v1
- Date: Tue, 10 Nov 2020 23:03:50 GMT
- Title: From Unsupervised Machine Translation To Adversarial Text Generation
- Authors: Ahmad Rashid, Alan Do-Omri, Md. Akmal Haidar, Qun Liu and Mehdi
Rezagholizadeh
- Abstract summary: We present a self-attention based bilingual adversarial text generator (B-GAN) which can learn to generate text from the encoder representation of an unsupervised neural machine translation system.
B-GAN is able to generate a distributed latent space representation which can be paired with an attention based decoder to generate fluent sentences.
- Score: 35.762161773313515
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a self-attention based bilingual adversarial text generator
(B-GAN) which can learn to generate text from the encoder representation of an
unsupervised neural machine translation system. B-GAN is able to generate a
distributed latent space representation which can be paired with an attention
based decoder to generate fluent sentences. When trained on an encoder shared
between two languages and paired with the appropriate decoder, it can generate
sentences in either language. B-GAN is trained using a combination of
reconstruction loss for auto-encoder, a cross domain loss for translation and a
GAN based adversarial loss for text generation. We demonstrate that B-GAN,
trained on monolingual corpora only using multiple losses, generates more
fluent sentences compared to monolingual baselines while effectively using half
the number of parameters.
Related papers
- Unsupervised Sign Language Translation and Generation [72.01216288379072]
We introduce an unsupervised sign language translation and generation network (USLNet)
USLNet learns from abundant single-modality (text and video) data without parallel sign language data.
We propose a sliding window method to address the issues of aligning variable-length text with video sequences.
arXiv Detail & Related papers (2024-02-12T15:39:05Z) - GanLM: Encoder-Decoder Pre-training with an Auxiliary Discriminator [114.8954615026781]
We propose a GAN-style model for encoder-decoder pre-training by introducing an auxiliary discriminator.
GanLM is trained with two pre-training objectives: replaced token detection and replaced token denoising.
Experiments in language generation benchmarks show that GanLM with the powerful language understanding capability outperforms various strong pre-trained language models.
arXiv Detail & Related papers (2022-12-20T12:51:11Z) - Improvement in Machine Translation with Generative Adversarial Networks [0.9612136532344103]
We take inspiration from RelGAN, a model for text generation, and NMT-GAN, an adversarial machine translation model, to implement a model that learns to transform awkward, non-fluent English sentences to fluent ones.
We utilize a parameter $lambda$ to control the amount of deviation from the input sentence, i.e. a trade-off between keeping the original tokens and modifying it to be more fluent.
arXiv Detail & Related papers (2021-11-30T06:51:13Z) - DeltaLM: Encoder-Decoder Pre-training for Language Generation and
Translation by Augmenting Pretrained Multilingual Encoders [92.90543340071007]
We introduce DeltaLM, a pretrained multilingual encoder-decoder model.
Specifically, we augment the pretrained multilingual encoder with a decoder and pre-train it in a self-supervised way.
Experiments show that DeltaLM outperforms various strong baselines on both natural language generation and translation tasks.
arXiv Detail & Related papers (2021-06-25T16:12:10Z) - Zero-Shot Cross-lingual Semantic Parsing [56.95036511882921]
We study cross-lingual semantic parsing as a zero-shot problem without parallel data for 7 test languages.
We propose a multi-task encoder-decoder model to transfer parsing knowledge to additional languages using only English-Logical form paired data.
Our system frames zero-shot parsing as a latent-space alignment problem and finds that pre-trained models can be improved to generate logical forms with minimal cross-lingual transfer penalty.
arXiv Detail & Related papers (2021-04-15T16:08:43Z) - Bi-Decoder Augmented Network for Neural Machine Translation [108.3931242633331]
We propose a novel Bi-Decoder Augmented Network (BiDAN) for the neural machine translation task.
Since each decoder transforms the representations of the input text into its corresponding language, jointly training with two target ends can make the shared encoder has the potential to produce a language-independent semantic space.
arXiv Detail & Related papers (2020-01-14T02:05:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.