Ranking Clarification Questions via Natural Language Inference
- URL: http://arxiv.org/abs/2008.07688v1
- Date: Tue, 18 Aug 2020 01:32:29 GMT
- Title: Ranking Clarification Questions via Natural Language Inference
- Authors: Vaibhav Kumar and Vikas Raunak and Jamie Callan
- Abstract summary: Given a natural language query, teaching machines to ask clarifying questions is of immense utility in practical natural language processing systems.
For the task of ranking clarification questions, we hypothesize that determining whether a clarification question pertains to a missing entry in a given post could be considered as a special case of Natural Language Inference (NLI)
We validate this hypothesis by incorporating representations from a Siamese BERT model fine-tuned on NLI and Multi-NLI datasets into our models.
- Score: 25.433933534561568
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Given a natural language query, teaching machines to ask clarifying questions
is of immense utility in practical natural language processing systems. Such
interactions could help in filling information gaps for better machine
comprehension of the query. For the task of ranking clarification questions, we
hypothesize that determining whether a clarification question pertains to a
missing entry in a given post (on QA forums such as StackExchange) could be
considered as a special case of Natural Language Inference (NLI), where both
the post and the most relevant clarification question point to a shared latent
piece of information or context. We validate this hypothesis by incorporating
representations from a Siamese BERT model fine-tuned on NLI and Multi-NLI
datasets into our models and demonstrate that our best performing model obtains
a relative performance improvement of 40 percent and 60 percent respectively
(on the key metric of Precision@1), over the state-of-the-art baseline(s) on
the two evaluation sets of the StackExchange dataset, thereby, significantly
surpassing the state-of-the-art.
Related papers
- Likelihood as a Performance Gauge for Retrieval-Augmented Generation [78.28197013467157]
We show that likelihoods serve as an effective gauge for language model performance.
We propose two methods that use question likelihood as a gauge for selecting and constructing prompts that lead to better performance.
arXiv Detail & Related papers (2024-11-12T13:14:09Z) - MST5 -- Multilingual Question Answering over Knowledge Graphs [1.6470999044938401]
Knowledge Graph Question Answering (KGQA) simplifies querying vast amounts of knowledge stored in a graph-based model using natural language.
Existing multilingual KGQA systems face challenges in achieving performance comparable to English systems.
We propose a simplified approach to enhance multilingual KGQA systems by incorporating linguistic context and entity information directly into the processing pipeline of a language model.
arXiv Detail & Related papers (2024-07-08T15:37:51Z) - Bridging the Gap: Deciphering Tabular Data Using Large Language Model [4.711941969101732]
This research marks the first application of large language models to table-based question answering tasks.
We have architected a distinctive module dedicated to the serialization of tables for seamless integration with expansive language models.
arXiv Detail & Related papers (2023-08-23T03:38:21Z) - Disco-Bench: A Discourse-Aware Evaluation Benchmark for Language
Modelling [70.23876429382969]
We propose a benchmark that can evaluate intra-sentence discourse properties across a diverse set of NLP tasks.
Disco-Bench consists of 9 document-level testsets in the literature domain, which contain rich discourse phenomena.
For linguistic analysis, we also design a diagnostic test suite that can examine whether the target models learn discourse knowledge.
arXiv Detail & Related papers (2023-07-16T15:18:25Z) - Extracting Multi-valued Relations from Language Models [36.944060044138304]
We analyze latent language representations for their potential to yield materialized multi-object relational knowledge.
For ranking candidate objects, we evaluate existing prompting techniques and propose new ones incorporating domain knowledge.
Among the selection methods, we find that choosing objects with a likelihood above a learned relation-specific threshold gives a 49.5% F1 score.
arXiv Detail & Related papers (2023-07-06T16:48:32Z) - Evaluating and Modeling Attribution for Cross-Lingual Question Answering [80.4807682093432]
This work is the first to study attribution for cross-lingual question answering.
We collect data in 5 languages to assess the attribution level of a state-of-the-art cross-lingual QA system.
We find that a substantial portion of the answers is not attributable to any retrieved passages.
arXiv Detail & Related papers (2023-05-23T17:57:46Z) - Syntax-informed Question Answering with Heterogeneous Graph Transformer [2.139714421848487]
We present a linguistics-informed question answering approach that extends and fine-tunes a pre-trained neural language model.
We illustrate the approach by the addition of syntactic information in the form of dependency and constituency graphic structures connecting tokens and virtual tokens.
arXiv Detail & Related papers (2022-04-01T07:48:03Z) - Automatically Identifying Semantic Bias in Crowdsourced Natural Language
Inference Datasets [78.6856732729301]
We introduce a model-driven, unsupervised technique to find "bias clusters" in a learned embedding space of hypotheses in NLI datasets.
interventions and additional rounds of labeling can be performed to ameliorate the semantic bias of the hypothesis distribution of a dataset.
arXiv Detail & Related papers (2021-12-16T22:49:01Z) - Reading Comprehension as Natural Language Inference: A Semantic Analysis [15.624486319943015]
We explore the utility of Natural language Inference (NLI) for Question Answering (QA)
We transform the one of the largest available MRC dataset (RACE) to an NLI form, and compare the performances of a state-of-the-art model (RoBERTa) on both forms.
We highlight clear categories for which the model is able to perform better when the data is presented in a coherent entailment form, and a structured question-answer concatenation form.
arXiv Detail & Related papers (2020-10-04T22:50:59Z) - ClarQ: A large-scale and diverse dataset for Clarification Question
Generation [67.1162903046619]
We devise a novel bootstrapping framework that assists in the creation of a diverse, large-scale dataset of clarification questions based on postcomments extracted from stackexchange.
We quantitatively demonstrate the utility of the newly created dataset by applying it to the downstream task of question-answering.
We release this dataset in order to foster research into the field of clarification question generation with the larger goal of enhancing dialog and question answering systems.
arXiv Detail & Related papers (2020-06-10T17:56:50Z) - Unsupervised Commonsense Question Answering with Self-Talk [71.63983121558843]
We propose an unsupervised framework based on self-talk as a novel alternative to commonsense tasks.
Inspired by inquiry-based discovery learning, our approach inquires language models with a number of information seeking questions.
Empirical results demonstrate that the self-talk procedure substantially improves the performance of zero-shot language model baselines.
arXiv Detail & Related papers (2020-04-11T20:43:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.