Self-explaining deep models with logic rule reasoning
- URL: http://arxiv.org/abs/2210.07024v2
- Date: Mon, 17 Oct 2022 08:24:42 GMT
- Title: Self-explaining deep models with logic rule reasoning
- Authors: Seungeon Lee, Xiting Wang, Sungwon Han, Xiaoyuan Yi, Xing Xie,
Meeyoung Cha
- Abstract summary: We present SELOR, a framework for integrating self-explaining capabilities into a given deep model.
By "human precision", we refer to the degree to which humans agree with the reasons models provide for their predictions.
- Score: 34.26828172603353
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present SELOR, a framework for integrating self-explaining capabilities
into a given deep model to achieve both high prediction performance and human
precision. By "human precision", we refer to the degree to which humans agree
with the reasons models provide for their predictions. Human precision affects
user trust and allows users to collaborate closely with the model. We
demonstrate that logic rule explanations naturally satisfy human precision with
the expressive power required for good predictive performance. We then
illustrate how to enable a deep model to predict and explain with logic rules.
Our method does not require predefined logic rule sets or human annotations and
can be learned efficiently and easily with widely-used deep learning modules in
a differentiable way. Extensive experiments show that our method gives
explanations closer to human decision logic than other methods while
maintaining the performance of deep learning models.
Related papers
- How Ambiguous are the Rationales for Natural Language Reasoning? A Simple Approach to Handling Rationale Uncertainty [0.0]
Rationales behind answers not only explain model decisions but boost language models to reason well on complex reasoning tasks.
It is non-trivial to estimate the degree to which the rationales are faithful enough to encourage model performance.
We propose how to deal with imperfect rationales causing aleatoric uncertainty.
arXiv Detail & Related papers (2024-02-22T07:12:34Z) - Evaluating the Utility of Model Explanations for Model Development [54.23538543168767]
We evaluate whether explanations can improve human decision-making in practical scenarios of machine learning model development.
To our surprise, we did not find evidence of significant improvement on tasks when users were provided with any of the saliency maps.
These findings suggest caution regarding the usefulness and potential for misunderstanding in saliency-based explanations.
arXiv Detail & Related papers (2023-12-10T23:13:23Z) - Longer Fixations, More Computation: Gaze-Guided Recurrent Neural
Networks [12.57650361978445]
Humans read texts at a varying pace, while machine learning models treat each token in the same way.
In this paper, we convert this intuition into a set of novel models with fixation-guided parallel RNNs or layers.
We find that, interestingly, the fixation duration predicted by neural networks bears some resemblance to humans' fixation.
arXiv Detail & Related papers (2023-10-31T21:32:11Z) - Emulating the Human Mind: A Neural-symbolic Link Prediction Model with
Fast and Slow Reasoning and Filtered Rules [4.979279893937017]
We introduce a novel Neural-Symbolic model named FaSt-FLiP.
Our objective is to combine a logical and neural model for enhanced link prediction.
arXiv Detail & Related papers (2023-10-21T12:45:11Z) - Human-Understandable Decision Making for Visual Recognition [30.30163407674527]
We propose a new framework to train a deep neural network by incorporating the prior of human perception into the model learning process.
The effectiveness of our proposed model is evaluated on two classical visual recognition tasks.
arXiv Detail & Related papers (2021-03-05T02:07:33Z) - Distilling Interpretable Models into Human-Readable Code [71.11328360614479]
Human-readability is an important and desirable standard for machine-learned model interpretability.
We propose to train interpretable models using conventional methods, and then distill them into concise, human-readable code.
We describe a piecewise-linear curve-fitting algorithm that produces high-quality results efficiently and reliably across a broad range of use cases.
arXiv Detail & Related papers (2021-01-21T01:46:36Z) - To what extent do human explanations of model behavior align with actual
model behavior? [91.67905128825402]
We investigated the extent to which human-generated explanations of models' inference decisions align with how models actually make these decisions.
We defined two alignment metrics that quantify how well natural language human explanations align with model sensitivity to input words.
We find that a model's alignment with human explanations is not predicted by the model's accuracy on NLI.
arXiv Detail & Related papers (2020-12-24T17:40:06Z) - Are Visual Explanations Useful? A Case Study in Model-in-the-Loop
Prediction [49.254162397086006]
We study explanations based on visual saliency in an image-based age prediction task.
We find that presenting model predictions improves human accuracy.
However, explanations of various kinds fail to significantly alter human accuracy or trust in the model.
arXiv Detail & Related papers (2020-07-23T20:39:40Z) - Leap-Of-Thought: Teaching Pre-Trained Models to Systematically Reason
Over Implicit Knowledge [96.92252296244233]
Large pre-trained language models (LMs) acquire some reasoning capacity, but this ability is difficult to control.
We show that LMs can be trained to reliably perform systematic reasoning combining both implicit, pre-trained knowledge and explicit natural language statements.
Our work paves a path towards open-domain systems that constantly improve by interacting with users who can instantly correct a model by adding simple natural language statements.
arXiv Detail & Related papers (2020-06-11T17:02:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.