Closed-Form Expressions for Global and Local Interpretation of Tsetlin
Machines with Applications to Explaining High-Dimensional Data
- URL: http://arxiv.org/abs/2007.13885v1
- Date: Mon, 27 Jul 2020 21:47:24 GMT
- Title: Closed-Form Expressions for Global and Local Interpretation of Tsetlin
Machines with Applications to Explaining High-Dimensional Data
- Authors: Christian D. Blakely, Ole-Christoffer Granmo
- Abstract summary: We propose closed-form expressions for understanding why a TM model makes a specific prediction (local interpretability)
We also introduce expressions for measuring the importance of feature value ranges for continuous features.
For both classification and regression, our evaluation show correspondence with SHAP as well as competitive prediction accuracy in comparison with XGBoost, Explainable Boosting Machines, and Neural Additive Models.
- Score: 7.05622249909585
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Tsetlin Machines (TMs) capture patterns using conjunctive clauses in
propositional logic, thus facilitating interpretation. However, recent TM-based
approaches mainly rely on inspecting the full range of clauses individually.
Such inspection does not necessarily scale to complex prediction problems that
require a large number of clauses. In this paper, we propose closed-form
expressions for understanding why a TM model makes a specific prediction (local
interpretability). Additionally, the expressions capture the most important
features of the model overall (global interpretability). We further introduce
expressions for measuring the importance of feature value ranges for continuous
features. The expressions are formulated directly from the conjunctive clauses
of the TM, making it possible to capture the role of features in real-time,
also during the learning process as the model evolves. Additionally, from the
closed-form expressions, we derive a novel data clustering algorithm for
visualizing high-dimensional data in three dimensions. Finally, we compare our
proposed approach against SHAP and state-of-the-art interpretable machine
learning techniques. For both classification and regression, our evaluation
show correspondence with SHAP as well as competitive prediction accuracy in
comparison with XGBoost, Explainable Boosting Machines, and Neural Additive
Models.
Related papers
- Pruning Literals for Highly Efficient Explainability at Word Level [13.249876381579158]
Tsetlin Machine(TM) is promising because of its capability of providing word-level explanation using proposition logic.
In this paper, we design a post-hoc pruning of clauses that eliminate the randomly placed literals in the clause.
Experiments on the publicly available YELP-HAT dataset demonstrate that the proposed pruned TM's attention map aligns more with the human attention map than the vanilla TM's attention map.
arXiv Detail & Related papers (2024-11-07T09:28:38Z) - Explaining Datasets in Words: Statistical Models with Natural Language Parameters [66.69456696878842]
We introduce a family of statistical models -- including clustering, time series, and classification models -- parameterized by natural language predicates.
We apply our framework to a wide range of problems: taxonomizing user chat dialogues, characterizing how they evolve across time, finding categories where one language model is better than the other.
arXiv Detail & Related papers (2024-09-13T01:40:20Z) - Exploring State Space and Reasoning by Elimination in Tsetlin Machines [14.150011713654331]
The Tsetlin Machine (TM) has gained significant attention in Machine Learning (ML)
TM is utilised to construct word embedding and describe target words using clauses.
To enhance the descriptive capacity of these clauses, we study the concept of Reasoning by Elimination (RbE) in clauses' formulation.
arXiv Detail & Related papers (2024-07-12T10:58:01Z) - When factorization meets argumentation: towards argumentative explanations [0.0]
We propose a novel model that combines factorization-based methods with argumentation frameworks (AFs)
Our framework seamlessly incorporates side information, such as user contexts, leading to more accurate predictions.
arXiv Detail & Related papers (2024-05-13T19:16:28Z) - TRIGO: Benchmarking Formal Mathematical Proof Reduction for Generative
Language Models [68.65075559137608]
We propose TRIGO, an ATP benchmark that not only requires a model to reduce a trigonometric expression with step-by-step proofs but also evaluates a generative LM's reasoning ability on formulas.
We gather trigonometric expressions and their reduced forms from the web, annotate the simplification process manually, and translate it into the Lean formal language system.
We develop an automatic generator based on Lean-Gym to create dataset splits of varying difficulties and distributions in order to thoroughly analyze the model's generalization ability.
arXiv Detail & Related papers (2023-10-16T08:42:39Z) - Fine-grained Retrieval Prompt Tuning [149.9071858259279]
Fine-grained Retrieval Prompt Tuning steers a frozen pre-trained model to perform the fine-grained retrieval task from the perspectives of sample prompt and feature adaptation.
Our FRPT with fewer learnable parameters achieves the state-of-the-art performance on three widely-used fine-grained datasets.
arXiv Detail & Related papers (2022-07-29T04:10:04Z) - Low-Rank Constraints for Fast Inference in Structured Models [110.38427965904266]
This work demonstrates a simple approach to reduce the computational and memory complexity of a large class of structured models.
Experiments with neural parameterized structured models for language modeling, polyphonic music modeling, unsupervised grammar induction, and video modeling show that our approach matches the accuracy of standard models at large state spaces.
arXiv Detail & Related papers (2022-01-08T00:47:50Z) - Locally Interpretable Model Agnostic Explanations using Gaussian
Processes [2.9189409618561966]
Local Interpretable Model-Agnostic Explanations (LIME) is a popular technique for explaining the prediction of a single instance.
We propose a Gaussian Process (GP) based variation of locally interpretable models.
We demonstrate that the proposed technique is able to generate faithful explanations using much fewer samples as compared to LIME.
arXiv Detail & Related papers (2021-08-16T05:49:01Z) - Did the Cat Drink the Coffee? Challenging Transformers with Generalized
Event Knowledge [59.22170796793179]
Transformers Language Models (TLMs) were tested on a benchmark for the textitdynamic estimation of thematic fit
Our results show that TLMs can reach performances that are comparable to those achieved by SDM.
However, additional analysis consistently suggests that TLMs do not capture important aspects of event knowledge.
arXiv Detail & Related papers (2021-07-22T20:52:26Z) - MAIRE -- A Model-Agnostic Interpretable Rule Extraction Procedure for
Explaining Classifiers [5.02231401459109]
The paper introduces a novel framework for extracting model-agnostic human interpretable rules to explain a classifier's output.
The framework is model agnostic, can be applied to any arbitrary classifier, and all types of attributes (including continuous, ordered, and unordered discrete)
arXiv Detail & Related papers (2020-11-03T06:53:06Z) - Understanding Neural Abstractive Summarization Models via Uncertainty [54.37665950633147]
seq2seq abstractive summarization models generate text in a free-form manner.
We study the entropy, or uncertainty, of the model's token-level predictions.
We show that uncertainty is a useful perspective for analyzing summarization and text generation models more broadly.
arXiv Detail & Related papers (2020-10-15T16:57:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.