Learning Semantically Meaningful Features for Interpretable
Classifications
- URL: http://arxiv.org/abs/2101.03919v1
- Date: Mon, 11 Jan 2021 14:35:16 GMT
- Title: Learning Semantically Meaningful Features for Interpretable
Classifications
- Authors: Sandareka Wickramanayake, Wynne Hsu, Mong Li Lee
- Abstract summary: SemCNN learns associations between visual features and word phrases.
Experiment results on multiple benchmark datasets demonstrate that SemCNN can learn features with clear semantic meaning.
- Score: 17.88784870849724
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning semantically meaningful features is important for Deep Neural
Networks to win end-user trust. Attempts to generate post-hoc explanations fall
short in gaining user confidence as they do not improve the interpretability of
feature representations learned by the models. In this work, we propose
Semantic Convolutional Neural Network (SemCNN) that has an additional Concept
layer to learn the associations between visual features and word phrases.
SemCNN employs an objective function that optimizes for both the prediction
accuracy as well as the semantic meaningfulness of the learned feature
representations. Further, SemCNN makes its decisions as a weighted sum of the
contributions of these features leading to fully interpretable decisions.
Experiment results on multiple benchmark datasets demonstrate that SemCNN can
learn features with clear semantic meaning and their corresponding
contributions to the model decision without compromising prediction accuracy.
Furthermore, these learned concepts are transferrable and can be applied to new
classes of objects that have similar concepts.
Related papers
- Manipulating Feature Visualizations with Gradient Slingshots [54.31109240020007]
We introduce a novel method for manipulating Feature Visualization (FV) without significantly impacting the model's decision-making process.
We evaluate the effectiveness of our method on several neural network models and demonstrate its capabilities to hide the functionality of arbitrarily chosen neurons.
arXiv Detail & Related papers (2024-01-11T18:57:17Z) - Seeing in Words: Learning to Classify through Language Bottlenecks [59.97827889540685]
Humans can explain their predictions using succinct and intuitive descriptions.
We show that a vision model whose feature representations are text can effectively classify ImageNet images.
arXiv Detail & Related papers (2023-06-29T00:24:42Z) - NxPlain: Web-based Tool for Discovery of Latent Concepts [16.446370662629555]
We present NxPlain, a web application that provides an explanation of a model's prediction using latent concepts.
NxPlain discovers latent concepts learned in a deep NLP model, provides an interpretation of the knowledge learned in the model, and explains its predictions based on the used concepts.
arXiv Detail & Related papers (2023-03-06T10:45:24Z) - TCNL: Transparent and Controllable Network Learning Via Embedding
Human-Guided Concepts [10.890006696574803]
We propose a novel method, Transparent and Controllable Network Learning (TCNL), to overcome such challenges.
Towards the goal of improving transparency-interpretability, in TCNL, we define some concepts for specific classification tasks through scientific human-intuition study.
We also build the concept mapper to visualize features extracted by the concept extractor in a human-intuitive way.
arXiv Detail & Related papers (2022-10-07T01:18:37Z) - Knowledge Enhanced Neural Networks for relational domains [83.9217787335878]
We focus on a specific method, KENN, a Neural-Symbolic architecture that injects prior logical knowledge into a neural network.
In this paper, we propose an extension of KENN for relational data.
arXiv Detail & Related papers (2022-05-31T13:00:34Z) - Self-Supervised Representation Learning on Neural Network Weights for
Model Characteristic Prediction [1.9659095632676094]
Self-Supervised Learning (SSL) has been shown to learn useful and information-preserving representations.
We propose to use SSL to learn neural representations of the weights of populations of Neural Networks (NNs)
Our empirical evaluation demonstrates that self-supervised representation learning in this domain is able to recover diverse NN model characteristics.
arXiv Detail & Related papers (2021-10-28T16:48:15Z) - Variational Structured Attention Networks for Deep Visual Representation
Learning [49.80498066480928]
We propose a unified deep framework to jointly learn both spatial attention maps and channel attention in a principled manner.
Specifically, we integrate the estimation and the interaction of the attentions within a probabilistic representation learning framework.
We implement the inference rules within the neural network, thus allowing for end-to-end learning of the probabilistic and the CNN front-end parameters.
arXiv Detail & Related papers (2021-03-05T07:37:24Z) - Explaining and Improving Model Behavior with k Nearest Neighbor
Representations [107.24850861390196]
We propose using k nearest neighbor representations to identify training examples responsible for a model's predictions.
We show that kNN representations are effective at uncovering learned spurious associations.
Our results indicate that the kNN approach makes the finetuned model more robust to adversarial inputs.
arXiv Detail & Related papers (2020-10-18T16:55:25Z) - Contextual Semantic Interpretability [16.18912769522768]
We look into semantic bottlenecks that capture context.
We use a two-layer semantic bottleneck that gathers attributes into interpretable, sparse groups.
Our model yields in predictions as accurate as a non-interpretable baseline when applied to a real-world test set of Flickr images.
arXiv Detail & Related papers (2020-09-18T09:47:05Z) - Neural Networks Enhancement with Logical Knowledge [83.9217787335878]
We propose an extension of KENN for relational data.
The results show that KENN is capable of increasing the performances of the underlying neural network even in the presence relational data.
arXiv Detail & Related papers (2020-09-13T21:12:20Z) - Making Sense of CNNs: Interpreting Deep Representations & Their
Invariances with INNs [19.398202091883366]
We present an approach based on INNs that (i) recovers the task-specific, learned invariances by disentangling the remaining factor of variation in the data and that (ii) invertibly transforms these invariances combined with the model representation into an equally expressive one with accessible semantic concepts.
Our invertible approach significantly extends the abilities to understand black box models by enabling post-hoc interpretations of state-of-the-art networks without compromising their performance.
arXiv Detail & Related papers (2020-08-04T19:27:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.