Verifying Properties of Tsetlin Machines
- URL: http://arxiv.org/abs/2303.14464v2
- Date: Sun, 2 Jul 2023 13:47:37 GMT
- Title: Verifying Properties of Tsetlin Machines
- Authors: Emilia Przybysz and Bimal Bhattarai and Cosimo Persia and Ana Ozaki
and Ole-Christoffer Granmo and Jivitesh Sharma
- Abstract summary: We present an exact encoding of TsMs into propositional logic and formally verify properties of TsMs using a SAT solver.
We consider notions of robustness and equivalence from the literature and adapt them for TsMs.
In our experiments, we employ the MNIST and IMDB datasets for (respectively) image and sentiment classification.
- Score: 18.870370171271126
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Tsetlin Machines (TsMs) are a promising and interpretable machine learning
method which can be applied for various classification tasks. We present an
exact encoding of TsMs into propositional logic and formally verify properties
of TsMs using a SAT solver. In particular, we introduce in this work a notion
of similarity of machine learning models and apply our notion to check for
similarity of TsMs. We also consider notions of robustness and equivalence from
the literature and adapt them for TsMs. Then, we show the correctness of our
encoding and provide results for the properties: adversarial robustness,
equivalence, and similarity of TsMs. In our experiments, we employ the MNIST
and IMDB datasets for (respectively) image and sentiment classification. We
discuss the results for verifying robustness obtained with TsMs with those in
the literature obtained with Binarized Neural Networks on MNIST.
Related papers
- Predicting Word Similarity in Context with Referential Translation Machines [0.0]
We identify the similarity between two words in English by casting the task as machine translation performance prediction (MTPP)
We use referential translation machines (RTMs) which allows a common representation of training and test sets.
RTMs can achieve the top results in Graded Word Similarity in Context (GWSC) task.
arXiv Detail & Related papers (2024-07-07T09:36:41Z) - Identifying Intensity of the Structure and Content in Tweets and the Discriminative Power of Attributes in Context with Referential Translation Machines [0.0]
We use referential translation machines (RTMs) to identify the similarity between an attribute and two words in English.
RTMs are also used to predict the intensity of the structure and content in tweets in English, Arabic, and Spanish.
arXiv Detail & Related papers (2024-07-06T18:58:10Z) - Which Syntactic Capabilities Are Statistically Learned by Masked
Language Models for Code? [51.29970742152668]
We highlight relying on accuracy-based measurements may lead to an overestimation of models' capabilities.
To address these issues, we introduce a technique called SyntaxEval in Syntactic Capabilities.
arXiv Detail & Related papers (2024-01-03T02:44:02Z) - Match me if you can: Semi-Supervised Semantic Correspondence Learning with Unpaired Images [76.47980643420375]
This paper builds on the hypothesis that there is an inherent data-hungry matter in learning semantic correspondences.
We demonstrate a simple machine annotator reliably enriches paired key points via machine supervision.
Our models surpass current state-of-the-art models on semantic correspondence learning benchmarks like SPair-71k, PF-PASCAL, and PF-WILLOW.
arXiv Detail & Related papers (2023-11-30T13:22:15Z) - Batch-Ensemble Stochastic Neural Networks for Out-of-Distribution
Detection [55.028065567756066]
Out-of-distribution (OOD) detection has recently received much attention from the machine learning community due to its importance in deploying machine learning models in real-world applications.
In this paper we propose an uncertainty quantification approach by modelling the distribution of features.
We incorporate an efficient ensemble mechanism, namely batch-ensemble, to construct the batch-ensemble neural networks (BE-SNNs) and overcome the feature collapse problem.
We show that BE-SNNs yield superior performance on several OOD benchmarks, such as the Two-Moons dataset, the FashionMNIST vs MNIST dataset, FashionM
arXiv Detail & Related papers (2022-06-26T16:00:22Z) - Towards Math-Aware Automated Classification and Similarity Search of
Scientific Publications: Methods of Mathematical Content Representations [0.456877715768796]
We investigate mathematical content representations suitable for the automated classification of and the similarity search in STEM documents.
The methods are evaluated on a subset of arXiv.org papers with the Mathematics Subject Classification (MSC) as a reference classification.
arXiv Detail & Related papers (2021-10-08T11:27:40Z) - More Than Words: Collocation Tokenization for Latent Dirichlet
Allocation Models [71.42030830910227]
We propose a new metric for measuring the clustering quality in settings where the models differ.
We show that topics trained with merged tokens result in topic keys that are clearer, more coherent, and more effective at distinguishing topics than those unmerged models.
arXiv Detail & Related papers (2021-08-24T14:08:19Z) - Interpretable and Low-Resource Entity Matching via Decoupling Feature
Learning from Decision Making [22.755892575582788]
Entity Matching aims at recognizing entity records that denote the same real-world object.
We propose a novel EM framework that consists of Heterogeneous Information Fusion (HIF) and Key Attribute Tree (KAT) Induction.
Our method is highly efficient and outperforms SOTA EM models in most cases.
arXiv Detail & Related papers (2021-06-08T08:27:31Z) - A Correspondence Variational Autoencoder for Unsupervised Acoustic Word
Embeddings [50.524054820564395]
We propose a new unsupervised model for mapping a variable-duration speech segment to a fixed-dimensional representation.
The resulting acoustic word embeddings can form the basis of search, discovery, and indexing systems for low- and zero-resource languages.
arXiv Detail & Related papers (2020-12-03T19:24:42Z) - Interpretable Multi-dataset Evaluation for Named Entity Recognition [110.64368106131062]
We present a general methodology for interpretable evaluation for the named entity recognition (NER) task.
The proposed evaluation method enables us to interpret the differences in models and datasets, as well as the interplay between them.
By making our analysis tool available, we make it easy for future researchers to run similar analyses and drive progress in this area.
arXiv Detail & Related papers (2020-11-13T10:53:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.