Tag-Set-Sequence Learning for Generating Question-Answer Pairs
- URL: http://arxiv.org/abs/2210.11608v1
- Date: Thu, 20 Oct 2022 21:51:00 GMT
- Title: Tag-Set-Sequence Learning for Generating Question-Answer Pairs
- Authors: Cheng Zhang, Jie Wang
- Abstract summary: We present a new method called tag-set sequence learning to tackle the problem of generating silly questions for texts.
We construct a system called TSS-Learner to learn tag-set sequences from given declarative sentences and the corresponding interrogative sentences.
We show that TSS-Learner can indeed generate adequate QAPs for certain texts that transformer-based models do poorly.
- Score: 10.48660454637293
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Transformer-based QG models can generate question-answer pairs (QAPs) with
high qualities, but may also generate silly questions for certain texts. We
present a new method called tag-set sequence learning to tackle this problem,
where a tag-set sequence is a sequence of tag sets to capture the syntactic and
semantic information of the underlying sentence, and a tag set consists of one
or more language feature tags, including, for example, semantic-role-labeling,
part-of-speech, named-entity-recognition, and sentiment-indication tags. We
construct a system called TSS-Learner to learn tag-set sequences from given
declarative sentences and the corresponding interrogative sentences, and derive
answers to the latter. We train a TSS-Learner model for the English language
using a small training dataset and show that it can indeed generate adequate
QAPs for certain texts that transformer-based models do poorly. Human
evaluation on the QAPs generated by TSS-Learner over SAT practice reading tests
is encouraging.
Related papers
- Constructing Cloze Questions Generatively [2.2719421441459406]
We present a generative method for constructing cloze questions from an article using neural networks and WordNet.
CQG selects an answer key for a given sentence, segments it into a sequence of instances, generates instance-level distractor candidates (IDCs) using a transformer and sibling synsets.
It then removes inappropriate IDCs, ranks the remaining IDCs based on contextual embedding similarities, as well as synset and lexical relatedness, forms distractor candidates by replacing instances with the corresponding top-ranked IDCs, and checks if they are legitimate phrases.
arXiv Detail & Related papers (2024-10-05T18:55:38Z) - SEMQA: Semi-Extractive Multi-Source Question Answering [94.04430035121136]
We introduce a new QA task for answering multi-answer questions by summarizing multiple diverse sources in a semi-extractive fashion.
We create the first dataset of this kind, QuoteSum, with human-written semi-extractive answers to natural and generated questions.
arXiv Detail & Related papers (2023-11-08T18:46:32Z) - SeqXGPT: Sentence-Level AI-Generated Text Detection [62.3792779440284]
We introduce a sentence-level detection challenge by synthesizing documents polished with large language models (LLMs)
We then propose textbfSequence textbfX (Check) textbfGPT, a novel method that utilizes log probability lists from white-box LLMs as features for sentence-level AIGT detection.
arXiv Detail & Related papers (2023-10-13T07:18:53Z) - Automatic Generation of Multiple-Choice Questions [7.310488568715925]
We present two methods to tackle the challenge of QAP generations.
A deep-learning-based end-to-end question generation system based on T5 Transformer with Preprocessing and Postprocessing Pipelines.
A sequence-learning-based scheme to generate adequate QAPs via meta-sequence representations of sentences.
arXiv Detail & Related papers (2023-03-25T22:45:54Z) - LabelPrompt: Effective Prompt-based Learning for Relation Classification [31.291466190218912]
This paper presents a novel prompt-based learning method, namely LabelPrompt, for the relation classification task.
Motivated by the intuition to GIVE MODEL CHOICES!'', we first define additional tokens to represent relation labels, which regard these tokens as the verbaliser with semantic initialisation.
Then, to mitigate inconsistency between predicted relations and given entities, we implement an entity-aware module with contrastive learning.
arXiv Detail & Related papers (2023-02-16T04:06:25Z) - Semantic Parsing for Conversational Question Answering over Knowledge
Graphs [63.939700311269156]
We develop a dataset where user questions are annotated with Sparql parses and system answers correspond to execution results thereof.
We present two different semantic parsing approaches and highlight the challenges of the task.
Our dataset and models are released at https://github.com/Edinburgh/SPICE.
arXiv Detail & Related papers (2023-01-28T14:45:11Z) - Prompting Language Models for Linguistic Structure [73.11488464916668]
We present a structured prompting approach for linguistic structured prediction tasks.
We evaluate this approach on part-of-speech tagging, named entity recognition, and sentence chunking.
We find that while PLMs contain significant prior knowledge of task labels due to task leakage into the pretraining corpus, structured prompting can also retrieve linguistic structure with arbitrary labels.
arXiv Detail & Related papers (2022-11-15T01:13:39Z) - Dict-TTS: Learning to Pronounce with Prior Dictionary Knowledge for
Text-to-Speech [88.22544315633687]
Polyphone disambiguation aims to capture accurate pronunciation knowledge from natural text sequences for reliable Text-to-speech systems.
We propose Dict-TTS, a semantic-aware generative text-to-speech model with an online website dictionary.
Experimental results in three languages show that our model outperforms several strong baseline models in terms of pronunciation accuracy.
arXiv Detail & Related papers (2022-06-05T10:50:34Z) - Generating Adequate Distractors for Multiple-Choice Questions [7.966913971277812]
Our method is a combination of part-of-speech tagging, named-entity tagging, semantic-role labeling, regular expressions, domain knowledge bases, word embeddings, word edit distance, WordNet, and other algorithms.
We show that, via experiments and by human judges, each MCQ has at least one adequate distractor and 84% of evaluations have three adequate distractors.
arXiv Detail & Related papers (2020-10-23T20:47:58Z) - Meta Sequence Learning for Generating Adequate Question-Answer Pairs [10.48660454637293]
We present a learning scheme to generate adequate QAPs via meta-sequence representations of sentences.
On a given declarative sentence, a trained MetaQA model converts it to a meta sequence, finds a matched MD, and uses the corresponding MIs and the input sentence to generate QAPs.
We show that MetaQA generates efficiently over the official SAT practice reading tests a large number of syntactically and semantically correct QAPs with over 97% accuracy.
arXiv Detail & Related papers (2020-10-04T16:28:13Z) - Syn-QG: Syntactic and Shallow Semantic Rules for Question Generation [49.671882751569534]
We develop SynQG, a set of transparent syntactic rules which transform declarative sentences into question-answer pairs.
We utilize PropBank argument descriptions and VerbNet state predicates to incorporate shallow semantic content.
In order to improve syntactic fluency and eliminate grammatically incorrect questions, we employ back-translation over the output of these syntactic rules.
arXiv Detail & Related papers (2020-04-18T19:57:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.