NeSyFOLD: Neurosymbolic Framework for Interpretable Image Classification
- URL: http://arxiv.org/abs/2301.12667v3
- Date: Sun, 20 Aug 2023 21:19:13 GMT
- Title: NeSyFOLD: Neurosymbolic Framework for Interpretable Image Classification
- Authors: Parth Padalkar, Huaduo Wang, Gopal Gupta
- Abstract summary: We present a novel framework called NeSyFOLD to create a neurosymbolic (NeSy) model for image classification tasks.
A rule-based machine learning algorithm called FOLD-SE-M is used to derive the stratified answer set program.
A justification for the predictions made by the NeSy model can be obtained using an ASP interpreter.
- Score: 1.3812010983144802
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Deep learning models such as CNNs have surpassed human performance in
computer vision tasks such as image classification. However, despite their
sophistication, these models lack interpretability which can lead to biased
outcomes reflecting existing prejudices in the data. We aim to make predictions
made by a CNN interpretable. Hence, we present a novel framework called
NeSyFOLD to create a neurosymbolic (NeSy) model for image classification tasks.
The model is a CNN with all layers following the last convolutional layer
replaced by a stratified answer set program (ASP). A rule-based machine
learning algorithm called FOLD-SE-M is used to derive the stratified answer set
program from binarized filter activations of the last convolutional layer. The
answer set program can be viewed as a rule-set, wherein the truth value of each
predicate depends on the activation of the corresponding kernel in the CNN. The
rule-set serves as a global explanation for the model and is interpretable. A
justification for the predictions made by the NeSy model can be obtained using
an ASP interpreter. We also use our NeSyFOLD framework with a CNN that is
trained using a sparse kernel learning technique called Elite BackProp (EBP).
This leads to a significant reduction in rule-set size without compromising
accuracy or fidelity thus improving scalability of the NeSy model and
interpretability of its rule-set. Evaluation is done on datasets with varied
complexity and sizes. To make the rule-set more intuitive to understand, we
propose a novel algorithm for labelling each kernel's corresponding predicate
in the rule-set with the semantic concept(s) it learns. We evaluate the
performance of our "semantic labelling algorithm" to quantify the efficacy of
the semantic labelling for both the NeSy model and the NeSy-EBP model.
Related papers
- Using Logic Programming and Kernel-Grouping for Improving
Interpretability of Convolutional Neural Networks [1.6317061277457001]
We present a neurosymbolic framework, NeSyFOLD-G that generates a symbolic rule-set using the last layer kernels of the CNN.
We show that grouping similar kernels leads to a significant reduction in the size of the rule-set generated by FOLD-SE-M.
We also propose a novel algorithm for labeling each predicate in the rule-set with the semantic concept(s) that its corresponding kernel group represents.
arXiv Detail & Related papers (2023-10-19T18:12:49Z) - Neural Additive Models for Location Scale and Shape: A Framework for
Interpretable Neural Regression Beyond the Mean [1.0923877073891446]
Deep neural networks (DNNs) have proven to be highly effective in a variety of tasks.
Despite this success, the inner workings of DNNs are often not transparent.
This lack of interpretability has led to increased research on inherently interpretable neural networks.
arXiv Detail & Related papers (2023-01-27T17:06:13Z) - Large-Margin Representation Learning for Texture Classification [67.94823375350433]
This paper presents a novel approach combining convolutional layers (CLs) and large-margin metric learning for training supervised models on small datasets for texture classification.
The experimental results on texture and histopathologic image datasets have shown that the proposed approach achieves competitive accuracy with lower computational cost and faster convergence when compared to equivalent CNNs.
arXiv Detail & Related papers (2022-06-17T04:07:45Z) - Adaptive Convolutional Dictionary Network for CT Metal Artifact
Reduction [62.691996239590125]
We propose an adaptive convolutional dictionary network (ACDNet) for metal artifact reduction.
Our ACDNet can automatically learn the prior for artifact-free CT images via training data and adaptively adjust the representation kernels for each input CT image.
Our method inherits the clear interpretability of model-based methods and maintains the powerful representation ability of learning-based methods.
arXiv Detail & Related papers (2022-05-16T06:49:36Z) - Do We Really Need a Learnable Classifier at the End of Deep Neural
Network? [118.18554882199676]
We study the potential of learning a neural network for classification with the classifier randomly as an ETF and fixed during training.
Our experimental results show that our method is able to achieve similar performances on image classification for balanced datasets.
arXiv Detail & Related papers (2022-03-17T04:34:28Z) - Dynamic Inference with Neural Interpreters [72.90231306252007]
We present Neural Interpreters, an architecture that factorizes inference in a self-attention network as a system of modules.
inputs to the model are routed through a sequence of functions in a way that is end-to-end learned.
We show that Neural Interpreters perform on par with the vision transformer using fewer parameters, while being transferrable to a new task in a sample efficient manner.
arXiv Detail & Related papers (2021-10-12T23:22:45Z) - Self-interpretable Convolutional Neural Networks for Text Classification [5.55878488884108]
This paper develops an approach for interpreting convolutional neural networks for text classification problems by exploiting the local-linear models inherent in ReLU-DNNs.
We show that our proposed technique produce parsimonious models that are self-interpretable and have comparable performance with respect to a more complex CNN model.
arXiv Detail & Related papers (2021-05-18T15:19:59Z) - Interpreting Graph Neural Networks for NLP With Differentiable Edge
Masking [63.49779304362376]
Graph neural networks (GNNs) have become a popular approach to integrating structural inductive biases into NLP models.
We introduce a post-hoc method for interpreting the predictions of GNNs which identifies unnecessary edges.
We show that we can drop a large proportion of edges without deteriorating the performance of the model.
arXiv Detail & Related papers (2020-10-01T17:51:19Z) - Approximation and Non-parametric Estimation of ResNet-type Convolutional
Neural Networks [52.972605601174955]
We show a ResNet-type CNN can attain the minimax optimal error rates in important function classes.
We derive approximation and estimation error rates of the aformentioned type of CNNs for the Barron and H"older classes.
arXiv Detail & Related papers (2019-03-24T19:42:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.