Interpretable Mixture of Experts
- URL: http://arxiv.org/abs/2206.02107v2
- Date: Thu, 25 May 2023 21:43:07 GMT
- Title: Interpretable Mixture of Experts
- Authors: Aya Abdelsalam Ismail, Sercan \"O. Arik, Jinsung Yoon, Ankur Taly,
Soheil Feizi and Tomas Pfister
- Abstract summary: Interpretable Mixture of Experts (IME) is an inherently-interpretable modeling framework.
IME is demonstrated to be more accurate than single interpretable models and perform comparably with existing state-of-the-art Deep Neural Networks (DNNs) in accuracy.
IME's explanations are compared to commonly-used post-hoc explanations methods through a user study.
- Score: 71.55701784196253
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The need for reliable model explanations is prominent for many machine
learning applications, particularly for tabular and time-series data as their
use cases often involve high-stakes decision making. Towards this goal, we
introduce a novel interpretable modeling framework, Interpretable Mixture of
Experts (IME), that yields high accuracy, comparable to `black-box' Deep Neural
Networks (DNNs) in many cases, along with useful interpretability capabilities.
IME consists of an assignment module and a mixture of experts, with each sample
being assigned to a single expert for prediction. We introduce multiple options
for IME based on the assignment and experts being interpretable. When the
experts are chosen to be interpretable such as linear models, IME yields an
inherently-interpretable architecture where the explanations produced by IME
are the exact descriptions of how the prediction is computed. In addition to
constituting a standalone inherently-interpretable architecture, IME has the
premise of being integrated with existing DNNs to offer interpretability to a
subset of samples while maintaining the accuracy of the DNNs. Through extensive
experiments on 15 tabular and time-series datasets, IME is demonstrated to be
more accurate than single interpretable models and perform comparably with
existing state-of-the-art DNNs in accuracy. On most datasets, IME even
outperforms DNNs, while providing faithful explanations. Lastly, IME's
explanations are compared to commonly-used post-hoc explanations methods
through a user study -- participants are able to better predict the model
behavior when given IME explanations, while finding IME's explanations more
faithful and trustworthy.
Related papers
- XForecast: Evaluating Natural Language Explanations for Time Series Forecasting [72.57427992446698]
Time series forecasting aids decision-making, especially for stakeholders who rely on accurate predictions.
Traditional explainable AI (XAI) methods, which underline feature or temporal importance, often require expert knowledge.
evaluating forecast NLEs is difficult due to the complex causal relationships in time series data.
arXiv Detail & Related papers (2024-10-18T05:16:39Z) - DISCRET: Synthesizing Faithful Explanations For Treatment Effect Estimation [21.172795461188578]
We propose DISCRET, a self-interpretable ITE framework that synthesizes faithful, rule-based explanations for each sample.
A key insight behind DISCRET is that explanations can serve dually as database queries to identify similar subgroups of samples.
We provide a novel RL algorithm to efficiently synthesize these explanations from a large search space.
arXiv Detail & Related papers (2024-06-02T04:01:08Z) - Explaining Text Similarity in Transformer Models [52.571158418102584]
Recent advances in explainable AI have made it possible to mitigate limitations by leveraging improved explanations for Transformers.
We use BiLRP, an extension developed for computing second-order explanations in bilinear similarity models, to investigate which feature interactions drive similarity in NLP models.
Our findings contribute to a deeper understanding of different semantic similarity tasks and models, highlighting how novel explainable AI methods enable in-depth analyses and corpus-level insights.
arXiv Detail & Related papers (2024-05-10T17:11:31Z) - Making Pre-trained Language Models Great on Tabular Prediction [50.70574370855663]
The transferability of deep neural networks (DNNs) has made significant progress in image and language processing.
We present TP-BERTa, a specifically pre-trained LM for tabular data prediction.
A novel relative magnitude tokenization converts scalar numerical feature values to finely discrete, high-dimensional tokens, and an intra-feature attention approach integrates feature values with the corresponding feature names.
arXiv Detail & Related papers (2024-03-04T08:38:56Z) - On Least Square Estimation in Softmax Gating Mixture of Experts [78.3687645289918]
We investigate the performance of the least squares estimators (LSE) under a deterministic MoE model.
We establish a condition called strong identifiability to characterize the convergence behavior of various types of expert functions.
Our findings have important practical implications for expert selection.
arXiv Detail & Related papers (2024-02-05T12:31:18Z) - InterpretCC: Intrinsic User-Centric Interpretability through Global Mixture of Experts [31.738009841932374]
Interpretability for neural networks is a trade-off between three key requirements.
We present InterpretCC, a family of interpretable-by-design neural networks that guarantee human-centric interpretability.
arXiv Detail & Related papers (2024-02-05T11:55:50Z) - Local Interpretable Model Agnostic Shap Explanations for machine
learning models [0.0]
We propose a methodology that we define as Local Interpretable Model Agnostic Shap Explanations (LIMASE)
This proposed technique uses Shapley values under the LIME paradigm to achieve the following (a) explain prediction of any model by using a locally faithful and interpretable decision tree model on which the Tree Explainer is used to calculate the shapley values and give visually interpretable explanations.
arXiv Detail & Related papers (2022-10-10T10:07:27Z) - Mixture of Linear Models Co-supervised by Deep Neural Networks [14.831346286039151]
We propose an approach to fill the gap between relatively simple explainable models and deep neural network (DNN) models.
Our main idea is a mixture of discriminative models that is trained with the guidance from a DNN.
arXiv Detail & Related papers (2021-08-05T02:08:35Z) - Masked Language Modeling and the Distributional Hypothesis: Order Word
Matters Pre-training for Little [74.49773960145681]
A possible explanation for the impressive performance of masked language model (MLM)-training is that such models have learned to represent the syntactic structures prevalent in NLP pipelines.
In this paper, we propose a different explanation: pre-trains succeed on downstream tasks almost entirely due to their ability to model higher-order word co-occurrence statistics.
Our results show that purely distributional information largely explains the success of pre-training, and underscore the importance of curating challenging evaluation datasets that require deeper linguistic knowledge.
arXiv Detail & Related papers (2021-04-14T06:30:36Z) - Explanations of Black-Box Model Predictions by Contextual Importance and
Utility [1.7188280334580195]
We present the Contextual Importance (CI) and Contextual Utility (CU) concepts to extract explanations easily understandable by experts as well as novice users.
This method explains the prediction results without transforming the model into an interpretable one.
We show the utility of explanations in car selection example and Iris flower classification by presenting complete (i.e. the causes of an individual prediction) and contrastive explanation.
arXiv Detail & Related papers (2020-05-30T06:49:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.