$\partial$-Explainer: Abductive Natural Language Inference via
Differentiable Convex Optimization
- URL: http://arxiv.org/abs/2105.03417v1
- Date: Fri, 7 May 2021 17:49:19 GMT
- Title: $\partial$-Explainer: Abductive Natural Language Inference via
Differentiable Convex Optimization
- Authors: Mokanarangan Thayaparan, Marco Valentino, Deborah Ferreira, Julia
Rozanova, Andr\'e Freitas
- Abstract summary: This paper presents a novel framework named $partial$-Explainer (Diff-Explainer) that combines the best of both worlds by casting the constrained optimization as part of a deep neural network.
Our experiments show up to $approx 10%$ improvement over non-differentiable solver while still providing explanations for supporting its inference.
- Score: 2.624902795082451
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Constrained optimization solvers with Integer Linear programming (ILP) have
been the cornerstone for explainable natural language inference during its
inception. ILP based approaches provide a way to encode explicit and
controllable assumptions casting natural language inference as an abductive
reasoning problem, where the solver constructs a plausible explanation for a
given hypothesis. While constrained based solvers provide explanations, they
are often limited by the use of explicit constraints and cannot be integrated
as part of broader deep neural architectures. In contrast, state-of-the-art
transformer-based models can learn from data and implicitly encode complex
constraints. However, these models are intrinsically black boxes. This paper
presents a novel framework named $\partial$-Explainer (Diff-Explainer) that
combines the best of both worlds by casting the constrained optimization as
part of a deep neural network via differentiable convex optimization and
fine-tuning pre-trained transformers for downstream explainable NLP tasks. To
demonstrate the efficacy of the framework, we transform the constraints
presented by TupleILP and integrate them with sentence embedding transformers
for the task of explainable science QA. Our experiments show up to $\approx
10\%$ improvement over non-differentiable solver while still providing
explanations for supporting its inference.
Related papers
- CF-OPT: Counterfactual Explanations for Structured Prediction [47.36059095502583]
optimization layers in deep neural networks have enjoyed a growing popularity in structured learning, improving the state of the art on a variety of applications.
Yet, these pipelines lack interpretability since they are made of two opaque layers: a highly non-linear prediction model, such as a deep neural network, and an optimization layer, which is typically a complex black-box solver.
Our goal is to improve the transparency of such methods by providing counterfactual explanations.
arXiv Detail & Related papers (2024-05-28T15:48:27Z) - Explaining Text Similarity in Transformer Models [52.571158418102584]
Recent advances in explainable AI have made it possible to mitigate limitations by leveraging improved explanations for Transformers.
We use BiLRP, an extension developed for computing second-order explanations in bilinear similarity models, to investigate which feature interactions drive similarity in NLP models.
Our findings contribute to a deeper understanding of different semantic similarity tasks and models, highlighting how novel explainable AI methods enable in-depth analyses and corpus-level insights.
arXiv Detail & Related papers (2024-05-10T17:11:31Z) - A Differentiable Integer Linear Programming Solver for Explanation-Based Natural Language Inference [17.467900115986158]
We introduce Diff-Comb Explainer, a neuro-symbolic architecture for explanation-based Natural Language Inference (NLI)
Diff-Comb Explainer does not necessitate a continuous relaxation of the semantic constraints, enabling a direct, more precise, and efficient incorporation of neural representations into the ILP formulation.
Our experiments demonstrate that Diff-Comb Explainer achieves superior performance when compared to conventional ILP solvers, neuro-symbolic black-box solvers, and Transformer-based encoders.
arXiv Detail & Related papers (2024-04-03T10:29:06Z) - Stable Nonconvex-Nonconcave Training via Linear Interpolation [51.668052890249726]
This paper presents a theoretical analysis of linearahead as a principled method for stabilizing (large-scale) neural network training.
We argue that instabilities in the optimization process are often caused by the nonmonotonicity of the loss landscape and show how linear can help by leveraging the theory of nonexpansive operators.
arXiv Detail & Related papers (2023-10-20T12:45:12Z) - Going Beyond Approximation: Encoding Constraints for Explainable
Multi-hop Inference via Differentiable Combinatorial Solvers [4.726777092009554]
Linear Programming (ILP) provides a viable mechanism to encode explicit and controllable assumptions about explainable multi-hop inference with natural language.
An ILP formulation is non-differentiable and cannot be integrated into broader deep learning architectures.
Diff-Comb Explainer demonstrates improved accuracy and explainability over non-differentiable solvers, Transformers and existing differentiable constraint-based multi-hop inference frameworks.
arXiv Detail & Related papers (2022-08-05T18:07:53Z) - Quasi Black-Box Variational Inference with Natural Gradients for
Bayesian Learning [84.90242084523565]
We develop an optimization algorithm suitable for Bayesian learning in complex models.
Our approach relies on natural gradient updates within a general black-box framework for efficient training with limited model-specific derivations.
arXiv Detail & Related papers (2022-05-23T18:54:27Z) - Hybrid Autoregressive Solver for Scalable Abductive Natural Language
Inference [2.867517731896504]
We propose a hybrid abductive solver that autoregressively combines a dense bi-encoder with a sparse model of explanatory power.
Our experiments demonstrate that the proposed framework can achieve performance comparable with the state-of-the-art cross-encoder.
arXiv Detail & Related papers (2021-07-25T19:29:53Z) - Causality-based Counterfactual Explanation for Classification Models [11.108866104714627]
We propose a prototype-based counterfactual explanation framework (ProCE)
ProCE is capable of preserving the causal relationship underlying the features of the counterfactual data.
In addition, we design a novel gradient-free optimization based on the multi-objective genetic algorithm that generates the counterfactual explanations.
arXiv Detail & Related papers (2021-05-03T09:25:59Z) - SML: a new Semantic Embedding Alignment Transformer for efficient
cross-lingual Natural Language Inference [71.57324258813674]
The ability of Transformers to perform with precision a variety of tasks such as question answering, Natural Language Inference (NLI) or summarising, have enable them to be ranked as one of the best paradigms to address this kind of tasks at present.
NLI is one of the best scenarios to test these architectures, due to the knowledge required to understand complex sentences and established a relation between a hypothesis and a premise.
In this paper, we propose a new architecture, siamese multilingual transformer, to efficiently align multilingual embeddings for Natural Language Inference.
arXiv Detail & Related papers (2021-03-17T13:23:53Z) - Efficient Semi-Implicit Variational Inference [65.07058307271329]
We propose an efficient and scalable semi-implicit extrapolational (SIVI)
Our method maps SIVI's evidence to a rigorous inference of lower gradient values.
arXiv Detail & Related papers (2021-01-15T11:39:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.