Abductive Commonsense Reasoning Exploiting Mutually Exclusive
Explanations
- URL: http://arxiv.org/abs/2305.14618v1
- Date: Wed, 24 May 2023 01:35:10 GMT
- Title: Abductive Commonsense Reasoning Exploiting Mutually Exclusive
Explanations
- Authors: Wenting Zhao and Justin T. Chiu and Claire Cardie and Alexander M.
Rush
- Abstract summary: Abductive reasoning aims to find plausible explanations for an event.
Existing approaches for abductive reasoning in natural language processing often rely on manually generated annotations for supervision.
This work proposes an approach for abductive commonsense reasoning that exploits the fact that only a subset of explanations is correct for a given context.
- Score: 118.0818807474809
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Abductive reasoning aims to find plausible explanations for an event. This
style of reasoning is critical for commonsense tasks where there are often
multiple plausible explanations. Existing approaches for abductive reasoning in
natural language processing (NLP) often rely on manually generated annotations
for supervision; however, such annotations can be subjective and biased.
Instead of using direct supervision, this work proposes an approach for
abductive commonsense reasoning that exploits the fact that only a subset of
explanations is correct for a given context. The method uses posterior
regularization to enforce a mutual exclusion constraint, encouraging the model
to learn the distinction between fluent explanations and plausible ones. We
evaluate our approach on a diverse set of abductive reasoning datasets;
experimental results show that our approach outperforms or is comparable to
directly applying pretrained language models in a zero-shot manner and other
knowledge-augmented zero-shot methods.
Related papers
- UNcommonsense Reasoning: Abductive Reasoning about Uncommon Situations [62.71847873326847]
We investigate the ability to model unusual, unexpected, and unlikely situations.
Given a piece of context with an unexpected outcome, this task requires reasoning abductively to generate an explanation.
We release a new English language corpus called UNcommonsense.
arXiv Detail & Related papers (2023-11-14T19:00:55Z) - Implicit Chain of Thought Reasoning via Knowledge Distillation [58.80851216530288]
Instead of explicitly producing the chain of thought reasoning steps, we use the language model's internal hidden states to perform implicit reasoning.
We find that this approach enables solving tasks previously not solvable without explicit chain-of-thought, at a speed comparable to no chain-of-thought.
arXiv Detail & Related papers (2023-11-02T17:59:49Z) - Phenomenal Yet Puzzling: Testing Inductive Reasoning Capabilities of Language Models with Hypothesis Refinement [92.61557711360652]
Language models (LMs) often fall short on inductive reasoning, despite achieving impressive success on research benchmarks.
We conduct a systematic study of the inductive reasoning capabilities of LMs through iterative hypothesis refinement.
We reveal several discrepancies between the inductive reasoning processes of LMs and humans, shedding light on both the potentials and limitations of using LMs in inductive reasoning tasks.
arXiv Detail & Related papers (2023-10-12T17:51:10Z) - Deductive Verification of Chain-of-Thought Reasoning [22.79166959432764]
Large Language Models (LLMs) benefit from Chain-of-Thought prompting in performing various reasoning tasks.
While CoT allows models to produce more comprehensive reasoning processes, its emphasis on intermediate reasoning steps can inadvertently introduce hallucinations and accumulated errors.
We propose Natural Program, a natural language-based deductive reasoning format.
arXiv Detail & Related papers (2023-06-06T17:18:56Z) - HOP, UNION, GENERATE: Explainable Multi-hop Reasoning without Rationale
Supervision [118.0818807474809]
This work proposes a principled, probabilistic approach for training explainable multi-hop QA systems without rationale supervision.
Our approach performs multi-hop reasoning by explicitly modeling rationales as sets, enabling the model to capture interactions between documents and sentences within a document.
arXiv Detail & Related papers (2023-05-23T16:53:49Z) - Prompting Contrastive Explanations for Commonsense Reasoning Tasks [74.7346558082693]
Large pretrained language models (PLMs) can achieve near-human performance on commonsense reasoning tasks.
We show how to use these same models to generate human-interpretable evidence.
arXiv Detail & Related papers (2021-06-12T17:06:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.