End-to-End Synthetic Data Generation for Domain Adaptation of Question
Answering Systems
- URL: http://arxiv.org/abs/2010.06028v1
- Date: Mon, 12 Oct 2020 21:10:18 GMT
- Title: End-to-End Synthetic Data Generation for Domain Adaptation of Question
Answering Systems
- Authors: Siamak Shakeri, Cicero Nogueira dos Santos, Henry Zhu, Patrick Ng,
Feng Nan, Zhiguo Wang, Ramesh Nallapati, Bing Xiang
- Abstract summary: Our model comprises a single transformer-based encoder-decoder network that is trained end-to-end to generate both answers and questions.
In a nutshell, we feed a passage to the encoder and ask the decoder to generate a question and an answer token-by-token.
- Score: 34.927828428293864
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose an end-to-end approach for synthetic QA data generation. Our model
comprises a single transformer-based encoder-decoder network that is trained
end-to-end to generate both answers and questions. In a nutshell, we feed a
passage to the encoder and ask the decoder to generate a question and an answer
token-by-token. The likelihood produced in the generation process is used as a
filtering score, which avoids the need for a separate filtering model. Our
generator is trained by fine-tuning a pretrained LM using maximum likelihood
estimation. The experimental results indicate significant improvements in the
domain adaptation of QA models outperforming current state-of-the-art methods.
Related papers
- Predictive Maintenance Model Based on Anomaly Detection in Induction
Motors: A Machine Learning Approach Using Real-Time IoT Data [0.0]
In this work, we demonstrate a novel anomaly detection system on induction motors used in pumps, compressors, fans, and other industrial machines.
We use a combination of pre-processing techniques and machine learning (ML) models with a low computational cost.
arXiv Detail & Related papers (2023-10-15T18:43:45Z) - Complexity Matters: Rethinking the Latent Space for Generative Modeling [65.64763873078114]
In generative modeling, numerous successful approaches leverage a low-dimensional latent space, e.g., Stable Diffusion.
In this study, we aim to shed light on this under-explored topic by rethinking the latent space from the perspective of model complexity.
arXiv Detail & Related papers (2023-07-17T07:12:29Z) - String-based Molecule Generation via Multi-decoder VAE [56.465033997245776]
We investigate the problem of string-based molecular generation via variational autoencoders (VAEs)
We propose a simple, yet effective idea to improve the performance of VAE for the task.
In our experiments, the proposed VAE model particularly performs well for generating a sample from out-of-domain distribution.
arXiv Detail & Related papers (2022-08-23T03:56:30Z) - CodeRL: Mastering Code Generation through Pretrained Models and Deep
Reinforcement Learning [92.36705236706678]
"CodeRL" is a new framework for program synthesis tasks through pretrained LMs and deep reinforcement learning.
During inference, we introduce a new generation procedure with a critical sampling strategy.
For the model backbones, we extended the encoder-decoder architecture of CodeT5 with enhanced learning objectives.
arXiv Detail & Related papers (2022-07-05T02:42:15Z) - Hyperdecoders: Instance-specific decoders for multi-task NLP [9.244884318445413]
We investigate input-conditioned hypernetworks for multi-tasking in NLP.
We generate parameter-efficient adaptations for a decoder using a hypernetwork conditioned on the output of an encoder.
arXiv Detail & Related papers (2022-03-15T22:39:53Z) - Disentangling Autoencoders (DAE) [0.0]
We propose a novel framework for autoencoders based on the principles of symmetry transformations in group-theory.
We believe that this model leads a new field for disentanglement learning based on autoencoders without regularizers.
arXiv Detail & Related papers (2022-02-20T22:59:13Z) - Autoencoding Variational Autoencoder [56.05008520271406]
We study the implications of this behaviour on the learned representations and also the consequences of fixing it by introducing a notion of self consistency.
We show that encoders trained with our self-consistency approach lead to representations that are robust (insensitive) to perturbations in the input introduced by adversarial attacks.
arXiv Detail & Related papers (2020-12-07T14:16:14Z) - Cross-Thought for Sentence Encoder Pre-training [89.32270059777025]
Cross-Thought is a novel approach to pre-training sequence encoder.
We train a Transformer-based sequence encoder over a large set of short sequences.
Experiments on question answering and textual entailment tasks demonstrate that our pre-trained encoder can outperform state-of-the-art encoders.
arXiv Detail & Related papers (2020-10-07T21:02:41Z) - Recent Developments Combining Ensemble Smoother and Deep Generative
Networks for Facies History Matching [58.720142291102135]
This research project focuses on the use of autoencoders networks to construct a continuous parameterization for facies models.
We benchmark seven different formulations, including VAE, generative adversarial network (GAN), Wasserstein GAN, variational auto-encoding GAN, principal component analysis (PCA) with cycle GAN, PCA with transfer style network and VAE with style loss.
arXiv Detail & Related papers (2020-05-08T21:32:42Z) - Batch norm with entropic regularization turns deterministic autoencoders
into generative models [14.65554816300632]
The variational autoencoder is a well defined deep generative model.
We show in this work that utilizing batch normalization as a source for non-determinism suffices to turn deterministic autoencoders into generative models.
arXiv Detail & Related papers (2020-02-25T02:42:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.