Weakly Supervised Explainable Phrasal Reasoning with Neural Fuzzy Logic
- URL: http://arxiv.org/abs/2109.08927v1
- Date: Sat, 18 Sep 2021 13:04:23 GMT
- Title: Weakly Supervised Explainable Phrasal Reasoning with Neural Fuzzy Logic
- Authors: Zijun Wu, Atharva Naik, Zi Xuan Zhang, Lili Mou
- Abstract summary: Natural language inference aims to determine the logical relationship between two sentences among the target labels Entailment, Contradiction, and Neutral.
Deep learning models have become a prevailing approach to NLI, but they lack interpretability and explainability.
In this work, we address the explainability for NLI by weakly supervised logical reasoning.
- Score: 24.868479255640718
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Natural language inference (NLI) aims to determine the logical relationship
between two sentences among the target labels Entailment, Contradiction, and
Neutral. In recent years, deep learning models have become a prevailing
approach to NLI, but they lack interpretability and explainability. In this
work, we address the explainability for NLI by weakly supervised logical
reasoning, and propose an Explainable Phrasal Reasoning (EPR) approach. Our
model first detects phrases as the semantic unit and aligns corresponding
phrases. Then, the model predicts the NLI label for the aligned phrases, and
induces the sentence label by fuzzy logic formulas. Our EPR is almost
everywhere differentiable and thus the system can be trained end-to-end in a
weakly supervised manner. We annotated a corpus and developed a set of metrics
to evaluate phrasal reasoning. Results show that our EPR yields much more
meaningful explanations in terms of F scores than previous studies. To the best
of our knowledge, we are the first to develop a weakly supervised phrasal
reasoning model for the NLI task.
Related papers
- LogicBench: Towards Systematic Evaluation of Logical Reasoning Ability of Large Language Models [52.03659714625452]
Recently developed large language models (LLMs) have been shown to perform remarkably well on a wide range of language understanding tasks.
But, can they really "reason" over the natural language?
This question has been receiving significant research attention and many reasoning skills such as commonsense, numerical, and qualitative have been studied.
arXiv Detail & Related papers (2024-04-23T21:08:49Z) - LogicAsker: Evaluating and Improving the Logical Reasoning Ability of Large Language Models [63.14196038655506]
We introduce LogicAsker, a novel approach for evaluating and enhancing the logical reasoning capabilities of large language models (LLMs)
Our methodology reveals significant gaps in LLMs' learning of logical rules, with identified reasoning failures ranging from 29% to 90% across different models.
We leverage these findings to construct targeted demonstration examples and fine-tune data, notably enhancing logical reasoning in models like GPT-4o by up to 5%.
arXiv Detail & Related papers (2024-01-01T13:53:53Z) - On the Interplay between Fairness and Explainability [28.37896468795247]
We perform a first study to understand how fairness and explainability influence each other.
We fine-tune pre-trained language models with several methods for bias mitigation.
We find that bias mitigation algorithms do not always lead to fairer models.
arXiv Detail & Related papers (2023-10-25T12:59:51Z) - Towards LogiGLUE: A Brief Survey and A Benchmark for Analyzing Logical Reasoning Capabilities of Language Models [56.34029644009297]
Large language models (LLMs) have demonstrated the ability to overcome various limitations of formal Knowledge Representation (KR) systems.
LLMs excel most in abductive reasoning, followed by deductive reasoning, while they are least effective at inductive reasoning.
We study single-task training, multi-task training, and "chain-of-thought" knowledge distillation fine-tuning technique to assess the performance of model.
arXiv Detail & Related papers (2023-10-02T01:00:50Z) - Learning Deductive Reasoning from Synthetic Corpus based on Formal Logic [14.503982715625902]
We study a synthetic corpus based approach for language models (LMs) to acquire logical deductive reasoning ability.
We adopt a well-grounded set of deduction rules based on formal logic theory, which can derive any other deduction rules when combined in a multistep way.
We empirically verify that LMs trained on FLD corpora acquire more generalizable reasoning ability.
arXiv Detail & Related papers (2023-08-11T13:15:35Z) - NELLIE: A Neuro-Symbolic Inference Engine for Grounded, Compositional, and Explainable Reasoning [59.16962123636579]
This paper proposes a new take on Prolog-based inference engines.
We replace handcrafted rules with a combination of neural language modeling, guided generation, and semi dense retrieval.
Our implementation, NELLIE, is the first system to demonstrate fully interpretable, end-to-end grounded QA.
arXiv Detail & Related papers (2022-09-16T00:54:44Z) - FOLIO: Natural Language Reasoning with First-Order Logic [147.50480350846726]
We present FOLIO, a human-annotated, logically complex and diverse dataset for reasoning in natural language (NL)
FOLIO consists of 1,430 examples (unique conclusions), each paired with one of 487 sets of premises used to deductively reason for the validity of each conclusion.
For both NL reasoning and NL-FOL translation, we benchmark multiple state-of-the-art language models.
arXiv Detail & Related papers (2022-09-02T06:50:11Z) - Logical Satisfiability of Counterfactuals for Faithful Explanations in
NLI [60.142926537264714]
We introduce the methodology of Faithfulness-through-Counterfactuals.
It generates a counterfactual hypothesis based on the logical predicates expressed in the explanation.
It then evaluates if the model's prediction on the counterfactual is consistent with that expressed logic.
arXiv Detail & Related papers (2022-05-25T03:40:59Z) - NILE : Natural Language Inference with Faithful Natural Language
Explanations [10.074153632701952]
We propose Natural-language Inference over Label-specific Explanations (NILE)
NILE is a novel NLI method which utilizes auto-generated label-specific explanations to produce labels along with its faithful explanation.
We discuss the faithfulness of NILE's explanations in terms of sensitivity of the decisions to the corresponding explanations.
arXiv Detail & Related papers (2020-05-25T13:56:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.