Interpretable Sequence Classification via Discrete Optimization
- URL: http://arxiv.org/abs/2010.02819v1
- Date: Tue, 6 Oct 2020 15:31:07 GMT
- Title: Interpretable Sequence Classification via Discrete Optimization
- Authors: Maayan Shvo, Andrew C. Li, Rodrigo Toro Icarte, Sheila A. McIlraith
- Abstract summary: In many applications such as healthcare monitoring or intrusion detection, early classification is crucial to prompt intervention.
In this work, we learn sequence classifiers that favour early classification from an evolving observation trace.
Our classifiers are interpretable---supporting explanation, counterfactual reasoning, and human-in-the-loop modification.
- Score: 26.899228003677138
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Sequence classification is the task of predicting a class label given a
sequence of observations. In many applications such as healthcare monitoring or
intrusion detection, early classification is crucial to prompt intervention. In
this work, we learn sequence classifiers that favour early classification from
an evolving observation trace. While many state-of-the-art sequence classifiers
are neural networks, and in particular LSTMs, our classifiers take the form of
finite state automata and are learned via discrete optimization. Our
automata-based classifiers are interpretable---supporting explanation,
counterfactual reasoning, and human-in-the-loop modification---and have strong
empirical performance. Experiments over a suite of goal recognition and
behaviour classification datasets show our learned automata-based classifiers
to have comparable test performance to LSTM-based classifiers, with the added
advantage of being interpretable.
Related papers
- Classifier Chain Networks for Multi-Label Classification [0.0]
The classifier chain is a widely used method for analyzing multi-labeled data sets.
We introduce a generalization of the chain: the classifier chain network.
arXiv Detail & Related papers (2024-11-04T21:56:13Z) - Anomaly Detection using Ensemble Classification and Evidence Theory [62.997667081978825]
We present a novel approach for novel detection using ensemble classification and evidence theory.
A pool selection strategy is presented to build a solid ensemble classifier.
We use uncertainty for the anomaly detection approach.
arXiv Detail & Related papers (2022-12-23T00:50:41Z) - Parametric Classification for Generalized Category Discovery: A Baseline
Study [70.73212959385387]
Generalized Category Discovery (GCD) aims to discover novel categories in unlabelled datasets using knowledge learned from labelled samples.
We investigate the failure of parametric classifiers, verify the effectiveness of previous design choices when high-quality supervision is available, and identify unreliable pseudo-labels as a key problem.
We propose a simple yet effective parametric classification method that benefits from entropy regularisation, achieves state-of-the-art performance on multiple GCD benchmarks and shows strong robustness to unknown class numbers.
arXiv Detail & Related papers (2022-11-21T18:47:11Z) - Prototypical Classifier for Robust Class-Imbalanced Learning [64.96088324684683]
We propose textitPrototypical, which does not require fitting additional parameters given the embedding network.
Prototypical produces balanced and comparable predictions for all classes even though the training set is class-imbalanced.
We test our method on CIFAR-10LT, CIFAR-100LT and Webvision datasets, observing that Prototypical obtains substaintial improvements compared with state of the arts.
arXiv Detail & Related papers (2021-10-22T01:55:01Z) - When in Doubt: Improving Classification Performance with Alternating
Normalization [57.39356691967766]
We introduce Classification with Alternating Normalization (CAN), a non-parametric post-processing step for classification.
CAN improves classification accuracy for challenging examples by re-adjusting their predicted class probability distribution.
We empirically demonstrate its effectiveness across a diverse set of classification tasks.
arXiv Detail & Related papers (2021-09-28T02:55:42Z) - Binary Classification from Multiple Unlabeled Datasets via Surrogate Set
Classification [94.55805516167369]
We propose a new approach for binary classification from m U-sets for $mge2$.
Our key idea is to consider an auxiliary classification task called surrogate set classification (SSC)
arXiv Detail & Related papers (2021-02-01T07:36:38Z) - Learning and Evaluating Representations for Deep One-class
Classification [59.095144932794646]
We present a two-stage framework for deep one-class classification.
We first learn self-supervised representations from one-class data, and then build one-class classifiers on learned representations.
In experiments, we demonstrate state-of-the-art performance on visual domain one-class classification benchmarks.
arXiv Detail & Related papers (2020-11-04T23:33:41Z) - Predicting Classification Accuracy When Adding New Unobserved Classes [8.325327265120283]
We study how a classifier's performance can be used to extrapolate its expected accuracy on a larger, unobserved set of classes.
We formulate a robust neural-network-based algorithm, "CleaneX", which learns to estimate the accuracy of such classifiers on arbitrarily large sets of classes.
arXiv Detail & Related papers (2020-10-28T14:37:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.