ArT: All-round Thinker for Unsupervised Commonsense Question-Answering
- URL: http://arxiv.org/abs/2112.13428v1
- Date: Sun, 26 Dec 2021 18:06:44 GMT
- Title: ArT: All-round Thinker for Unsupervised Commonsense Question-Answering
- Authors: Jiawei Wang and Hai Zhao
- Abstract summary: We propose an approach of All-round Thinker (ArT) by fully taking association during knowledge generating.
We evaluate it on three commonsense QA benchmarks: COPA, SocialIQA and SCT.
- Score: 54.068032948300655
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Without labeled question-answer pairs for necessary training, unsupervised
commonsense question-answering (QA) appears to be extremely challenging due to
its indispensable unique prerequisite on commonsense source like knowledge
bases (KBs), which are usually highly resource consuming in construction.
Recently pre-trained language models (PrLMs) show effectiveness as an
alternative for commonsense clues when they play a role of knowledge generator.
However, existing work simply generates hundreds of pseudo-answers, or roughly
performs knowledge generation according to templates once for all, which may
result in much noise and thus hinders the quality of generated knowledge.
Motivated by human thinking experience, we propose an approach of All-round
Thinker (ArT) by fully taking association during knowledge generating. In
detail, our model first focuses on key parts in the given context, and then
generates highly related knowledge on such a basis in an association way like
human thinking. Besides, for casual reasoning, a reverse thinking mechanism is
proposed to conduct bidirectional inferring between cause and effect. ArT is
totally unsupervised and KBs-free. We evaluate it on three commonsense QA
benchmarks: COPA, SocialIQA and SCT. On all scales of PrLM backbones, ArT shows
its brilliant performance and outperforms previous advanced unsupervised
models.
Related papers
- Multi-hop Commonsense Knowledge Injection Framework for Zero-Shot
Commonsense Question Answering [6.086719709100659]
We propose a novel multi-hop commonsense knowledge injection framework.
Our framework achieves state-of-art performance on five commonsense question answering benchmarks.
arXiv Detail & Related papers (2023-05-10T07:13:47Z) - CIKQA: Learning Commonsense Inference with a Unified
Knowledge-in-the-loop QA Paradigm [120.98789964518562]
We argue that due to the large scale of commonsense knowledge, it is infeasible to annotate a large enough training set for each task to cover all commonsense for learning.
We focus on investigating models' commonsense inference capabilities from two perspectives.
We name the benchmark as Commonsense Inference with Knowledge-in-the-loop Question Answering (CIKQA)
arXiv Detail & Related papers (2022-10-12T14:32:39Z) - Rainier: Reinforced Knowledge Introspector for Commonsense Question
Answering [74.90418840431425]
We present Rainier, or Reinforced Knowledge Introspector, that learns to generate contextually relevant knowledge in response to given questions.
Our approach starts by imitating knowledge generated by GPT-3, then learns to generate its own knowledge via reinforcement learning.
Our work is the first to report that knowledge generated by models that are orders of magnitude smaller than GPT-3, even without direct supervision on the knowledge itself, can exceed the quality of knowledge elicited from GPT-3 for commonsense QA.
arXiv Detail & Related papers (2022-10-06T17:34:06Z) - A Unified End-to-End Retriever-Reader Framework for Knowledge-based VQA [67.75989848202343]
This paper presents a unified end-to-end retriever-reader framework towards knowledge-based VQA.
We shed light on the multi-modal implicit knowledge from vision-language pre-training models to mine its potential in knowledge reasoning.
Our scheme is able to not only provide guidance for knowledge retrieval, but also drop these instances potentially error-prone towards question answering.
arXiv Detail & Related papers (2022-06-30T02:35:04Z) - elBERto: Self-supervised Commonsense Learning for Question Answering [131.51059870970616]
We propose a Self-supervised Bidirectional Representation Learning of Commonsense framework, which is compatible with off-the-shelf QA model architectures.
The framework comprises five self-supervised tasks to force the model to fully exploit the additional training signals from contexts containing rich commonsense.
elBERto achieves substantial improvements on out-of-paragraph and no-effect questions where simple lexical similarity comparison does not help.
arXiv Detail & Related papers (2022-03-17T16:23:45Z) - Enhancing Question Generation with Commonsense Knowledge [33.289599417096206]
We propose a multi-task learning framework to introduce commonsense knowledge into question generation process.
Experimental results on SQuAD show that our proposed methods are able to noticeably improve the QG performance on both automatic and human evaluation metrics.
arXiv Detail & Related papers (2021-06-19T08:58:13Z) - Unsupervised Commonsense Question Answering with Self-Talk [71.63983121558843]
We propose an unsupervised framework based on self-talk as a novel alternative to commonsense tasks.
Inspired by inquiry-based discovery learning, our approach inquires language models with a number of information seeking questions.
Empirical results demonstrate that the self-talk procedure substantially improves the performance of zero-shot language model baselines.
arXiv Detail & Related papers (2020-04-11T20:43:37Z) - What Does My QA Model Know? Devising Controlled Probes using Expert
Knowledge [36.13528043657398]
We investigate whether state-of-the-art QA models have general knowledge about word definitions and general taxonomic reasoning.
We use a methodology for automatically building datasets from various types of expert knowledge.
Our evaluation confirms that transformer-based QA models are already predisposed to recognize certain types of structural lexical knowledge.
arXiv Detail & Related papers (2019-12-31T15:05:54Z) - Teaching Pretrained Models with Commonsense Reasoning: A Preliminary
KB-Based Approach [24.954288132238293]
We propose a method to teach pretrained models with commonsense reasoning by leveraging the structured knowledge in ConceptNet.
Experimental results demonstrate that, when refined on these training examples, the pretrained models consistently improve their performance on tasks that require commonsense reasoning.
arXiv Detail & Related papers (2019-09-20T23:58:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.