Implicit Mixture of Interpretable Experts for Global and Local
Interpretability
- URL: http://arxiv.org/abs/2212.00471v1
- Date: Thu, 1 Dec 2022 12:54:42 GMT
- Title: Implicit Mixture of Interpretable Experts for Global and Local
Interpretability
- Authors: Nathan Elazar, Kerry Taylor
- Abstract summary: We investigate the feasibility of using mixtures of interpretable experts (MoIE) to build interpretable image classifiers on MNIST10.
We find that a naively trained MoIE will learn to 'cheat', whereby the black-box router will solve the classification problem by itself.
We propose a novel implicit parameterization scheme that allows us to build mixtures of arbitrary numbers of experts.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: We investigate the feasibility of using mixtures of interpretable experts
(MoIE) to build interpretable image classifiers on MNIST10. MoIE uses a
black-box router to assign each input to one of many inherently interpretable
experts, thereby providing insight into why a particular classification
decision was made. We find that a naively trained MoIE will learn to 'cheat',
whereby the black-box router will solve the classification problem by itself,
with each expert simply learning a constant function for one particular class.
We propose to solve this problem by introducing interpretable routers and
training the black-box router's decisions to match the interpretable router. In
addition, we propose a novel implicit parameterization scheme that allows us to
build mixtures of arbitrary numbers of experts, allowing us to study how
classification performance, local and global interpretability vary as the
number of experts is increased. Our new model, dubbed Implicit Mixture of
Interpretable Experts (IMoIE) can match state-of-the-art classification
accuracy on MNIST10 while providing local interpretability, and can provide
global interpretability albeit at the cost of reduced classification accuracy.
Related papers
- Monet: Mixture of Monosemantic Experts for Transformers [33.8311330578753]
We introduce Mixture of Monosemantic Experts for Transformers (Monet) architecture.
Monet incorporates sparse dictionary learning directly into end-to-end Mixture-of-Experts pretraining.
Our analyses demonstrate mutual exclusivity of knowledge across experts and showcase the parametric knowledge encapsulated within individual experts.
arXiv Detail & Related papers (2024-12-05T13:06:03Z) - Beyond Parameter Count: Implicit Bias in Soft Mixture of Experts [44.09546603624385]
We introduce a notion of expert specialization for Soft MoE.
We show that when there are many small experts, the architecture is implicitly biased in a fashion that allows us to efficiently approximate the specialized expert subset.
arXiv Detail & Related papers (2024-09-02T00:39:00Z) - Generalization Error Analysis for Sparse Mixture-of-Experts: A Preliminary Study [65.11303133775857]
Mixture-of-Experts (MoE) computation amalgamates predictions from several specialized sub-models (referred to as experts)
Sparse MoE selectively engages only a limited number, or even just one expert, significantly reducing overhead while empirically preserving, and sometimes even enhancing, performance.
arXiv Detail & Related papers (2024-03-26T05:48:02Z) - Merge, Then Compress: Demystify Efficient SMoE with Hints from Its Routing Policy [84.11508381847929]
Sparsely activated Mixture-of-Experts (SMoE) has shown promise to scale up the learning capacity of neural networks.
We propose M-SMoE, which leverages routing statistics to guide expert merging.
Our MC-SMoE achieves up to 80% memory and a 20% FLOPs reduction, with virtually no loss in performance.
arXiv Detail & Related papers (2023-10-02T16:51:32Z) - MoEC: Mixture of Expert Clusters [93.63738535295866]
Sparsely Mixture of Experts (MoE) has received great interest due to its promising scaling capability with affordable computational overhead.
MoE converts dense layers into sparse experts, and utilizes a gated routing network to make experts conditionally activated.
However, as the number of experts grows, MoE with outrageous parameters suffers from overfitting and sparse data allocation.
arXiv Detail & Related papers (2022-07-19T06:09:55Z) - Interpretable Mixture of Experts [71.55701784196253]
Interpretable Mixture of Experts (IME) is an inherently-interpretable modeling framework.
IME is demonstrated to be more accurate than single interpretable models and perform comparably with existing state-of-the-art Deep Neural Networks (DNNs) in accuracy.
IME's explanations are compared to commonly-used post-hoc explanations methods through a user study.
arXiv Detail & Related papers (2022-06-05T06:40:15Z) - Sparsely-gated Mixture-of-Expert Layers for CNN Interpretability [3.021134753248103]
Sparsely-gated Mixture of Expert (MoE) layers have been successfully applied for scaling large transformers.
In this work, we apply sparse MoE layers to CNNs for computer vision tasks and analyze the resulting effect on model interpretability.
arXiv Detail & Related papers (2022-04-22T09:40:23Z) - On the Representation Collapse of Sparse Mixture of Experts [102.83396489230375]
Sparse mixture of experts provides larger model capacity while requiring a constant computational overhead.
It employs the routing mechanism to distribute input tokens to the best-matched experts according to their hidden representations.
However, learning such a routing mechanism encourages token clustering around expert centroids, implying a trend toward representation collapse.
arXiv Detail & Related papers (2022-04-20T01:40:19Z) - DEMix Layers: Disentangling Domains for Modular Language Modeling [92.57761975953453]
We introduce a new domain expert mixture (DEMix) layer that enables conditioning a language model (LM) on the domain of the input text.
A DEMix layer is a collection of expert feedforward networks, each specialized to a domain.
Experiments show that DEMix layers reduce test-time perplexity, increase training efficiency, and enable rapid adaptation with little overhead.
arXiv Detail & Related papers (2021-08-11T05:15:33Z) - Making Neural Networks Interpretable with Attribution: Application to
Implicit Signals Prediction [11.427019313283997]
We propose a novel formulation of interpretable deep neural networks for the attribution task.
Using masked weights, hidden features can be deeply attributed, split into several input-restricted sub-networks and trained as a boosted mixture of experts.
arXiv Detail & Related papers (2020-08-26T06:46:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.