Learning as Abduction: Trainable Natural Logic Theorem Prover for
Natural Language Inference
- URL: http://arxiv.org/abs/2010.15909v2
- Date: Tue, 1 Dec 2020 12:02:44 GMT
- Title: Learning as Abduction: Trainable Natural Logic Theorem Prover for
Natural Language Inference
- Authors: Lasha Abzianidze
- Abstract summary: We implement a learning method in a theorem prover for natural language.
We show that it improves the performance of the theorem prover on the SICK dataset by 1.4% while still maintaining high precision.
The obtained results are competitive with the state of the art among logic-based systems.
- Score: 0.4962199635155534
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Tackling Natural Language Inference with a logic-based method is becoming
less and less common. While this might have been counterintuitive several
decades ago, nowadays it seems pretty obvious. The main reasons for such a
conception are that (a) logic-based methods are usually brittle when it comes
to processing wide-coverage texts, and (b) instead of automatically learning
from data, they require much of manual effort for development. We make a step
towards to overcome such shortcomings by modeling learning from data as
abduction: reversing a theorem-proving procedure to abduce semantic relations
that serve as the best explanation for the gold label of an inference problem.
In other words, instead of proving sentence-level inference relations with the
help of lexical relations, the lexical relations are proved taking into account
the sentence-level inference relations. We implement the learning method in a
tableau theorem prover for natural language and show that it improves the
performance of the theorem prover on the SICK dataset by 1.4% while still
maintaining high precision (>94%). The obtained results are competitive with
the state of the art among logic-based systems.
Related papers
- TabVer: Tabular Fact Verification with Natural Logic [11.002475880349452]
We propose a set-theoretic interpretation of numerals and arithmetic functions in the context of natural logic.
We leverage large language models to generate arithmetic expressions by generating questions about salient parts of a claim which are answered by executing functions on tables.
In a few-shot setting on FEVEROUS, we achieve an accuracy of 71.4, outperforming both fully neural and symbolic reasoning models by 3.4 points.
arXiv Detail & Related papers (2024-11-02T00:36:34Z) - Scaling Synthetic Logical Reasoning Datasets with Context-Sensitive Declarative Grammars [0.6537995248511139]
We present a declarative framework with flexible context-sensitive rules binding multiple languages.
We construct first-order logic problems by selecting up to 32 premises and one hypothesis.
We demonstrate that using semantic constraints during generation and careful English verbalization of predicates enhances logical reasoning without hurting natural English tasks.
arXiv Detail & Related papers (2024-06-16T18:10:49Z) - Enhancing Systematic Decompositional Natural Language Inference Using Informal Logic [51.967603572656266]
We introduce a consistent and theoretically grounded approach to annotating decompositional entailment.
We find that our new dataset, RDTE, has a substantially higher internal consistency (+9%) than prior decompositional entailment datasets.
We also find that training an RDTE-oriented entailment classifier via knowledge distillation and employing it in an entailment tree reasoning engine significantly improves both accuracy and proof quality.
arXiv Detail & Related papers (2024-02-22T18:55:17Z) - LINC: A Neurosymbolic Approach for Logical Reasoning by Combining
Language Models with First-Order Logic Provers [60.009969929857704]
Logical reasoning is an important task for artificial intelligence with potential impacts on science, mathematics, and society.
In this work, we reformulating such tasks as modular neurosymbolic programming, which we call LINC.
We observe significant performance gains on FOLIO and a balanced subset of ProofWriter for three different models in nearly all experimental conditions we evaluate.
arXiv Detail & Related papers (2023-10-23T17:58:40Z) - QA-NatVer: Question Answering for Natural Logic-based Fact Verification [11.002475880349452]
We propose to use question answering to predict natural logic operators.
In a few-shot setting on FEVER, our approach outperforms the best baseline by $4.3$ accuracy points.
A human evaluation indicates that our approach produces more plausible with fewer erroneous natural logic operators than previous natural logic-based systems.
arXiv Detail & Related papers (2023-10-22T06:27:31Z) - Modeling Hierarchical Reasoning Chains by Linking Discourse Units and
Key Phrases for Reading Comprehension [80.99865844249106]
We propose a holistic graph network (HGN) which deals with context at both discourse level and word level, as the basis for logical reasoning.
Specifically, node-level and type-level relations, which can be interpreted as bridges in the reasoning process, are modeled by a hierarchical interaction mechanism.
arXiv Detail & Related papers (2023-06-21T07:34:27Z) - ChatABL: Abductive Learning via Natural Language Interaction with
ChatGPT [72.83383437501577]
Large language models (LLMs) have recently demonstrated significant potential in mathematical abilities.
LLMs currently have difficulty in bridging perception, language understanding and reasoning capabilities.
This paper presents a novel method for integrating LLMs into the abductive learning framework.
arXiv Detail & Related papers (2023-04-21T16:23:47Z) - Maieutic Prompting: Logically Consistent Reasoning with Recursive
Explanations [71.2950434944196]
We develop Maieutic Prompting, which infers a correct answer to a question even from the noisy and inconsistent generations of language models.
Maieutic Prompting achieves up to 20% better accuracy than state-of-the-art prompting methods.
arXiv Detail & Related papers (2022-05-24T06:36:42Z) - On the Paradox of Learning to Reason from Data [86.13662838603761]
We show that BERT can attain near-perfect accuracy on in-distribution test examples while failing to generalize to other data distributions over the exact same problem space.
Our study provides an explanation for this paradox: instead of learning to emulate the correct reasoning function, BERT has in fact learned statistical features that inherently exist in logical reasoning problems.
arXiv Detail & Related papers (2022-05-23T17:56:48Z) - Non-ground Abductive Logic Programming with Probabilistic Integrity
Constraints [2.624902795082451]
In this paper, we consider a richer logic language, coping with probabilistic abduction with variables.
We first present the overall abductive language, and its semantics according to the Distribution Semantics.
We then introduce a proof procedure, obtained by extending one previously presented, and prove its soundness and completeness.
arXiv Detail & Related papers (2021-08-06T10:22:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.