Learning to Diversify for Product Question Generation
- URL: http://arxiv.org/abs/2207.02534v1
- Date: Wed, 6 Jul 2022 09:26:41 GMT
- Title: Learning to Diversify for Product Question Generation
- Authors: Haggai Roitman, Uriel Singer, Yotam Eshel, Alexander Nus, Eliyahu
Kiperwasser
- Abstract summary: We show how the T5 pre-trained Transformer encoder-decoder model can be fine-tuned for the task.
We propose a novel learning-to-diversify (LTD) fine-tuning approach that allows to enrich the language learned by the underlying Transformer model.
- Score: 68.69526529887607
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: We address the product question generation task. For a given product
description, our goal is to generate questions that reflect potential user
information needs that are either missing or not well covered in the
description. Moreover, we wish to cover diverse user information needs that may
span a multitude of product types. To this end, we first show how the T5
pre-trained Transformer encoder-decoder model can be fine-tuned for the task.
Yet, while the T5 generated questions have a reasonable quality compared to the
state-of-the-art method for the task (KPCNet), many of such questions are still
too general, resulting in a sub-optimal global question diversity. As an
alternative, we propose a novel learning-to-diversify (LTD) fine-tuning
approach that allows to enrich the language learned by the underlying
Transformer model. Our empirical evaluation shows that, using our approach
significantly improves the global diversity of the underlying Transformer
model, while preserves, as much as possible, its generation relevance.
Related papers
- In-Context Learning with Representations: Contextual Generalization of Trained Transformers [66.78052387054593]
In-context learning (ICL) refers to a capability of pretrained large language models, which can learn a new task given a few examples during inference.
This paper investigates the training dynamics of transformers by gradient descent through the lens of non-linear regression tasks.
arXiv Detail & Related papers (2024-08-19T16:47:46Z) - Prompt-Engineering and Transformer-based Question Generation and
Evaluation [0.0]
This paper aims to find the best method to generate questions from textual data through a transformer model and prompt engineering.
The generated questions were compared against the baseline questions in the SQuAD dataset to evaluate the effectiveness of four different prompts.
arXiv Detail & Related papers (2023-10-29T01:45:30Z) - VOLTA: Improving Generative Diversity by Variational Mutual Information Maximizing Autoencoder [38.35049378875308]
We introduce VOLTA, a framework that elevates generative diversity by bridging Transformer with VAE.
We perform comprehensive experiments with two types of Transformers on six datasets to show that our approach can significantly improve generative diversity while maintaining generative quality.
arXiv Detail & Related papers (2023-07-03T08:45:42Z) - A Comprehensive Survey on Applications of Transformers for Deep Learning
Tasks [60.38369406877899]
Transformer is a deep neural network that employs a self-attention mechanism to comprehend the contextual relationships within sequential data.
transformer models excel in handling long dependencies between input sequence elements and enable parallel processing.
Our survey encompasses the identification of the top five application domains for transformer-based models.
arXiv Detail & Related papers (2023-06-11T23:13:51Z) - An Overview on Controllable Text Generation via Variational
Auto-Encoders [15.97186478109836]
Recent advances in neural-based generative modeling have reignited the hopes of having computer systems capable of conversing with humans.
Latent variable models (LVM) such as variational auto-encoders (VAEs) are designed to characterize the distributional pattern of textual data.
This overview gives an introduction to existing generation schemes, problems associated with text variational auto-encoders, and a review of several applications about the controllable generation.
arXiv Detail & Related papers (2022-11-15T07:36:11Z) - Semantic-aware Modular Capsule Routing for Visual Question Answering [55.03883681191765]
We propose a Semantic-aware modUlar caPsulE framework, termed as SUPER, to better capture the instance-specific vision-semantic characteristics.
We comparatively justify the effectiveness and generalization ability of our proposed SUPER scheme over five benchmark datasets.
arXiv Detail & Related papers (2022-07-21T10:48:37Z) - KAT: A Knowledge Augmented Transformer for Vision-and-Language [56.716531169609915]
We propose a novel model - Knowledge Augmented Transformer (KAT) - which achieves a strong state-of-the-art result on the open-domain multimodal task of OK-VQA.
Our approach integrates implicit and explicit knowledge in an end to end encoder-decoder architecture, while still jointly reasoning over both knowledge sources during answer generation.
An additional benefit of explicit knowledge integration is seen in improved interpretability of model predictions in our analysis.
arXiv Detail & Related papers (2021-12-16T04:37:10Z) - Inquisitive Question Generation for High Level Text Comprehension [60.21497846332531]
We introduce INQUISITIVE, a dataset of 19K questions that are elicited while a person is reading through a document.
We show that readers engage in a series of pragmatic strategies to seek information.
We evaluate question generation models based on GPT-2 and show that our model is able to generate reasonable questions.
arXiv Detail & Related papers (2020-10-04T19:03:39Z) - Simplifying Paragraph-level Question Generation via Transformer Language
Models [0.0]
Question generation (QG) is a natural language generation task where a model is trained to ask questions corresponding to some input text.
A single Transformer-based unidirectional language model leveraging transfer learning can be used to produce high quality questions.
Our QG model, finetuned from GPT-2 Small, outperforms several paragraph-level QG baselines on the SQuAD dataset by 0.95 METEOR points.
arXiv Detail & Related papers (2020-05-03T14:57:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.