Background Knowledge Injection for Interpretable Sequence Classification
- URL: http://arxiv.org/abs/2006.14248v1
- Date: Thu, 25 Jun 2020 08:36:05 GMT
- Title: Background Knowledge Injection for Interpretable Sequence Classification
- Authors: Severin Gsponer, Luca Costabello, Chan Le Van, Sumit Pai, Christophe
Gueret, Georgiana Ifrim, Freddy Lecue
- Abstract summary: We introduce a novel sequence learning algorithm that balances predictive power and interpretability.
We extend the classic subsequence feature space with groups of symbols generated by background knowledge injected via word or graph embeddings.
We also present a new measure to evaluate the interpretability of a set of symbolic features based on the symbol embeddings.
- Score: 13.074542699823933
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Sequence classification is the supervised learning task of building models
that predict class labels of unseen sequences of symbols. Although accuracy is
paramount, in certain scenarios interpretability is a must. Unfortunately, such
trade-off is often hard to achieve since we lack human-independent
interpretability metrics. We introduce a novel sequence learning algorithm,
that combines (i) linear classifiers - which are known to strike a good balance
between predictive power and interpretability, and (ii) background knowledge
embeddings. We extend the classic subsequence feature space with groups of
symbols which are generated by background knowledge injected via word or graph
embeddings, and use this new feature space to learn a linear classifier. We
also present a new measure to evaluate the interpretability of a set of
symbolic features based on the symbol embeddings. Experiments on human activity
recognition from wearables and amino acid sequence classification show that our
classification approach preserves predictive power, while delivering more
interpretable models.
Related papers
- Simple and Interpretable Probabilistic Classifiers for Knowledge Graphs [0.0]
We describe an inductive approach based on learning simple belief networks.
We show how such models can be converted into (probabilistic) axioms (or rules)
arXiv Detail & Related papers (2024-07-09T17:05:52Z) - A Multi-Grained Self-Interpretable Symbolic-Neural Model For
Single/Multi-Labeled Text Classification [29.075766631810595]
We propose a Symbolic-Neural model that can learn to explicitly predict class labels of text spans from a constituency tree.
As the structured language model learns to predict constituency trees in a self-supervised manner, only raw texts and sentence-level labels are required as training data.
Our experiments demonstrate that our approach could achieve good prediction accuracy in downstream tasks.
arXiv Detail & Related papers (2023-03-06T03:25:43Z) - Generalization Bounds for Few-Shot Transfer Learning with Pretrained
Classifiers [26.844410679685424]
We study the ability of foundation models to learn representations for classification that are transferable to new, unseen classes.
We show that the few-shot error of the learned feature map on new classes is small in case of class-feature-variability collapse.
arXiv Detail & Related papers (2022-12-23T18:46:05Z) - Learning Debiased and Disentangled Representations for Semantic
Segmentation [52.35766945827972]
We propose a model-agnostic and training scheme for semantic segmentation.
By randomly eliminating certain class information in each training iteration, we effectively reduce feature dependencies among classes.
Models trained with our approach demonstrate strong results on multiple semantic segmentation benchmarks.
arXiv Detail & Related papers (2021-10-31T16:15:09Z) - Prototypical Classifier for Robust Class-Imbalanced Learning [64.96088324684683]
We propose textitPrototypical, which does not require fitting additional parameters given the embedding network.
Prototypical produces balanced and comparable predictions for all classes even though the training set is class-imbalanced.
We test our method on CIFAR-10LT, CIFAR-100LT and Webvision datasets, observing that Prototypical obtains substaintial improvements compared with state of the arts.
arXiv Detail & Related papers (2021-10-22T01:55:01Z) - Intersection Regularization for Extracting Semantic Attributes [72.53481390411173]
We consider the problem of supervised classification, such that the features that the network extracts match an unseen set of semantic attributes.
For example, when learning to classify images of birds into species, we would like to observe the emergence of features that zoologists use to classify birds.
We propose training a neural network with discrete top-level activations, which is followed by a multi-layered perceptron (MLP) and a parallel decision tree.
arXiv Detail & Related papers (2021-03-22T14:32:44Z) - Theoretical Insights Into Multiclass Classification: A High-dimensional
Asymptotic View [82.80085730891126]
We provide the first modernally precise analysis of linear multiclass classification.
Our analysis reveals that the classification accuracy is highly distribution-dependent.
The insights gained may pave the way for a precise understanding of other classification algorithms.
arXiv Detail & Related papers (2020-11-16T05:17:29Z) - Network Classifiers Based on Social Learning [71.86764107527812]
We propose a new way of combining independently trained classifiers over space and time.
The proposed architecture is able to improve prediction performance over time with unlabeled data.
We show that this strategy results in consistent learning with high probability, and it yields a robust structure against poorly trained classifiers.
arXiv Detail & Related papers (2020-10-23T11:18:20Z) - A Framework to Learn with Interpretation [2.3741312212138896]
We present a novel framework to jointly learn a predictive model and its associated interpretation model.
We seek for a small-size dictionary of high level attribute functions that take as inputs the outputs of selected hidden layers.
A detailed pipeline to visualize the learnt features is also developed.
arXiv Detail & Related papers (2020-10-19T09:26:28Z) - Interpretable Sequence Classification via Discrete Optimization [26.899228003677138]
In many applications such as healthcare monitoring or intrusion detection, early classification is crucial to prompt intervention.
In this work, we learn sequence classifiers that favour early classification from an evolving observation trace.
Our classifiers are interpretable---supporting explanation, counterfactual reasoning, and human-in-the-loop modification.
arXiv Detail & Related papers (2020-10-06T15:31:07Z) - Latent Embedding Feedback and Discriminative Features for Zero-Shot
Classification [139.44681304276]
zero-shot learning aims to classify unseen categories for which no data is available during training.
Generative Adrial Networks synthesize unseen class features by leveraging class-specific semantic embeddings.
We propose to enforce semantic consistency at all stages of zero-shot learning: training, feature synthesis and classification.
arXiv Detail & Related papers (2020-03-17T17:34:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.