Neural Unification for Logic Reasoning over Natural Language
- URL: http://arxiv.org/abs/2109.08460v1
- Date: Fri, 17 Sep 2021 10:48:39 GMT
- Title: Neural Unification for Logic Reasoning over Natural Language
- Authors: Gabriele Picco, Hoang Thanh Lam, Marco Luca Sbodio, Vanessa Lopez
Garcia
- Abstract summary: Automated Theorem Proving deals with the development of computer programs being able to show that some conjectures (queries) are a logical consequence of a set of axioms (facts and rules)
Recent approaches have proposed transformer-based architectures for deriving conjectures given axioms expressed in natural language (English)
In this work we propose a new architecture, namely the Neural Unifier, which achieves state-of-the-art results in term of generalisation.
- Score: 0.28675177318965034
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Automated Theorem Proving (ATP) deals with the development of computer
programs being able to show that some conjectures (queries) are a logical
consequence of a set of axioms (facts and rules). There exists several
successful ATPs where conjectures and axioms are formally provided (e.g.
formalised as First Order Logic formulas). Recent approaches, such as (Clark et
al., 2020), have proposed transformer-based architectures for deriving
conjectures given axioms expressed in natural language (English). The
conjecture is verified through a binary text classifier, where the transformers
model is trained to predict the truth value of a conjecture given the axioms.
The RuleTaker approach of (Clark et al., 2020) achieves appealing results both
in terms of accuracy and in the ability to generalize, showing that when the
model is trained with deep enough queries (at least 3 inference steps), the
transformers are able to correctly answer the majority of queries (97.6%) that
require up to 5 inference steps. In this work we propose a new architecture,
namely the Neural Unifier, and a relative training procedure, which achieves
state-of-the-art results in term of generalisation, showing that mimicking a
well-known inference procedure, the backward chaining, it is possible to answer
deep queries even when the model is trained only on shallow ones. The approach
is demonstrated in experiments using a diverse set of benchmark data.
Related papers
- LINC: A Neurosymbolic Approach for Logical Reasoning by Combining
Language Models with First-Order Logic Provers [60.009969929857704]
Logical reasoning is an important task for artificial intelligence with potential impacts on science, mathematics, and society.
In this work, we reformulating such tasks as modular neurosymbolic programming, which we call LINC.
We observe significant performance gains on FOLIO and a balanced subset of ProofWriter for three different models in nearly all experimental conditions we evaluate.
arXiv Detail & Related papers (2023-10-23T17:58:40Z) - QA-NatVer: Question Answering for Natural Logic-based Fact Verification [11.002475880349452]
We propose to use question answering to predict natural logic operators.
In a few-shot setting on FEVER, our approach outperforms the best baseline by $4.3$ accuracy points.
A human evaluation indicates that our approach produces more plausible with fewer erroneous natural logic operators than previous natural logic-based systems.
arXiv Detail & Related papers (2023-10-22T06:27:31Z) - TRIGO: Benchmarking Formal Mathematical Proof Reduction for Generative
Language Models [68.65075559137608]
We propose TRIGO, an ATP benchmark that not only requires a model to reduce a trigonometric expression with step-by-step proofs but also evaluates a generative LM's reasoning ability on formulas.
We gather trigonometric expressions and their reduced forms from the web, annotate the simplification process manually, and translate it into the Lean formal language system.
We develop an automatic generator based on Lean-Gym to create dataset splits of varying difficulties and distributions in order to thoroughly analyze the model's generalization ability.
arXiv Detail & Related papers (2023-10-16T08:42:39Z) - Dual Box Embeddings for the Description Logic EL++ [16.70961576041243]
Similar to Knowledge Graphs (KGs), Knowledge Graphs are often incomplete, and maintaining and constructing them has proved challenging.
Similar to KGs, a promising approach is to learn embeddings in a latent vector space, while additionally ensuring they adhere to the semantics of the underlying DL.
We propose a novel ontology embedding method named Box$2$EL for the DL EL++, which represents both concepts and roles as boxes.
arXiv Detail & Related papers (2023-01-26T14:13:37Z) - MetaLogic: Logical Reasoning Explanations with Fine-Grained Structure [129.8481568648651]
We propose a benchmark to investigate models' logical reasoning capabilities in complex real-life scenarios.
Based on the multi-hop chain of reasoning, the explanation form includes three main components.
We evaluate the current best models' performance on this new explanation form.
arXiv Detail & Related papers (2022-10-22T16:01:13Z) - Language Models Are Greedy Reasoners: A Systematic Formal Analysis of
Chain-of-Thought [10.524051272257614]
Large language models (LLMs) have shown remarkable reasoning capabilities given chain-of-thought prompts.
We present a new synthetic question-answering dataset called PrOntoQA, where each example is generated as a synthetic world model.
This allows us to parse the generated chain-of-thought into symbolic proofs for formal analysis.
arXiv Detail & Related papers (2022-10-03T21:34:32Z) - Non-ground Abductive Logic Programming with Probabilistic Integrity
Constraints [2.624902795082451]
In this paper, we consider a richer logic language, coping with probabilistic abduction with variables.
We first present the overall abductive language, and its semantics according to the Distribution Semantics.
We then introduce a proof procedure, obtained by extending one previously presented, and prove its soundness and completeness.
arXiv Detail & Related papers (2021-08-06T10:22:12Z) - PRover: Proof Generation for Interpretable Reasoning over Rules [81.40404921232192]
We propose a transformer-based model that answers binary questions over rule-bases and generates the corresponding proofs.
Our model learns to predict nodes and edges corresponding to proof graphs in an efficient constrained training paradigm.
We conduct experiments on synthetic, hand-authored, and human-paraphrased rule-bases to show promising results for QA and proof generation.
arXiv Detail & Related papers (2020-10-06T15:47:53Z) - Generative Language Modeling for Automated Theorem Proving [94.01137612934842]
This work is motivated by the possibility that a major limitation of automated theorem provers compared to humans might be addressable via generation from language models.
We present an automated prover and proof assistant, GPT-f, for the Metamath formalization language, and analyze its performance.
arXiv Detail & Related papers (2020-09-07T19:50:10Z) - Leap-Of-Thought: Teaching Pre-Trained Models to Systematically Reason
Over Implicit Knowledge [96.92252296244233]
Large pre-trained language models (LMs) acquire some reasoning capacity, but this ability is difficult to control.
We show that LMs can be trained to reliably perform systematic reasoning combining both implicit, pre-trained knowledge and explicit natural language statements.
Our work paves a path towards open-domain systems that constantly improve by interacting with users who can instantly correct a model by adding simple natural language statements.
arXiv Detail & Related papers (2020-06-11T17:02:20Z) - Transformers as Soft Reasoners over Language [33.291806251021185]
This paper investigates a problem where the facts and rules are provided as natural language sentences, thus bypassing a formal representation.
We train transformers to emulate reason (or reasoning) over these sentences using synthetically generated data.
Our models, that we call RuleTakers, provide the first empirical demonstration that this kind of soft reasoning over language is learnable.
arXiv Detail & Related papers (2020-02-14T04:23:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.