Interpretable Machine Learning based on Functional ANOVA Framework:
Algorithms and Comparisons
- URL: http://arxiv.org/abs/2305.15670v1
- Date: Thu, 25 May 2023 02:40:52 GMT
- Title: Interpretable Machine Learning based on Functional ANOVA Framework:
Algorithms and Comparisons
- Authors: Linwei Hu, Vijayan N. Nair, Agus Sudjianto, Aijun Zhang, and Jie Chen
- Abstract summary: In the early days of machine learning (ML), the emphasis was on developing complex algorithms to achieve best predictive performance.
Recently, researchers are compromising on small increases in predictive performance to develop algorithms that are inherently interpretable.
The paper proposes a new algorithm, called GAMI-Lin-T, that also uses trees like EBM, but it does linear fits instead of piecewise constants within the partitions.
- Score: 9.10422407200807
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the early days of machine learning (ML), the emphasis was on developing
complex algorithms to achieve best predictive performance. To understand and
explain the model results, one had to rely on post hoc explainability
techniques, which are known to have limitations. Recently, with the recognition
that interpretability is just as important, researchers are compromising on
small increases in predictive performance to develop algorithms that are
inherently interpretable. While doing so, the ML community has rediscovered the
use of low-order functional ANOVA (fANOVA) models that have been known in the
statistical literature for some time. This paper starts with a description of
challenges with post hoc explainability and reviews the fANOVA framework with a
focus on main effects and second-order interactions. This is followed by an
overview of two recently developed techniques: Explainable Boosting Machines or
EBM (Lou et al., 2013) and GAMI-Net (Yang et al., 2021b). The paper proposes a
new algorithm, called GAMI-Lin-T, that also uses trees like EBM, but it does
linear fits instead of piecewise constants within the partitions. There are
many other differences, including the development of a new interaction
filtering algorithm. Finally, the paper uses simulated and real datasets to
compare selected ML algorithms. The results show that GAMI-Lin-T and GAMI-Net
have comparable performances, and both are generally better than EBM.
Related papers
- Attribute-to-Delete: Machine Unlearning via Datamodel Matching [65.13151619119782]
Machine unlearning -- efficiently removing a small "forget set" training data on a pre-divertrained machine learning model -- has recently attracted interest.
Recent research shows that machine unlearning techniques do not hold up in such a challenging setting.
arXiv Detail & Related papers (2024-10-30T17:20:10Z) - Energy-based learning algorithms for analog computing: a comparative
study [2.0937431058291933]
Energy-based learning algorithms have recently gained a surge of interest due to their compatibility with analog hardware.
We compare seven learning algorithms, namely contrastive learning (CL), equilibrium propagation (EP) and coupled learning (CpL)
We find that negative perturbations are better than positive ones, and highlight the centered variant of EP as the best-performing algorithm.
arXiv Detail & Related papers (2023-12-22T22:49:58Z) - Performance Evaluation and Comparison of a New Regression Algorithm [4.125187280299247]
We compare the performance of a newly proposed regression algorithm against four conventional machine learning algorithms.
The reader is free to replicate our results since we have provided the source code in a GitHub repository.
arXiv Detail & Related papers (2023-06-15T13:01:16Z) - Improved Algorithms for Neural Active Learning [74.89097665112621]
We improve the theoretical and empirical performance of neural-network(NN)-based active learning algorithms for the non-parametric streaming setting.
We introduce two regret metrics by minimizing the population loss that are more suitable in active learning than the one used in state-of-the-art (SOTA) related work.
arXiv Detail & Related papers (2022-10-02T05:03:38Z) - Using Model-Based Trees with Boosting to Fit Low-Order Functional ANOVA
Models [5.131758478675364]
Low-order functional ANOVA models have been rediscovered in the machine learning (ML) community under the guise of inherently interpretable machine learning.
We propose a new algorithm, called GAMI-Tree, that is similar to EBM, but has a number of features that lead to better performance.
We use simulated and real datasets to compare the performance and interpretability of GAMI-Tree with EBM and GAMI-Net.
arXiv Detail & Related papers (2022-07-14T14:23:14Z) - Simple Stochastic and Online Gradient DescentAlgorithms for Pairwise
Learning [65.54757265434465]
Pairwise learning refers to learning tasks where the loss function depends on a pair instances.
Online descent (OGD) is a popular approach to handle streaming data in pairwise learning.
In this paper, we propose simple and online descent to methods for pairwise learning.
arXiv Detail & Related papers (2021-11-23T18:10:48Z) - Dual Optimization for Kolmogorov Model Learning Using Enhanced Gradient
Descent [8.714458129632158]
Kolmogorov model (KM) is an interpretable and predictable representation approach to learning the underlying probabilistic structure of a set of random variables.
We propose a computationally scalable KM learning algorithm, based on the regularized dual optimization combined with enhanced gradient descent (GD) method.
It is shown that the accuracy of logical relation mining for interpretability by using the proposed KM learning algorithm exceeds $80%$.
arXiv Detail & Related papers (2021-07-11T10:33:02Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z) - Learning outside the Black-Box: The pursuit of interpretable models [78.32475359554395]
This paper proposes an algorithm that produces a continuous global interpretation of any given continuous black-box function.
Our interpretation represents a leap forward from the previous state of the art.
arXiv Detail & Related papers (2020-11-17T12:39:44Z) - Surrogate Locally-Interpretable Models with Supervised Machine Learning
Algorithms [8.949704905866888]
Supervised Machine Learning algorithms have become popular in recent years due to their superior predictive performance over traditional statistical methods.
The main focus is on interpretability, the resulting surrogate model also has reasonably good predictive performance.
arXiv Detail & Related papers (2020-07-28T23:46:16Z) - MPLP++: Fast, Parallel Dual Block-Coordinate Ascent for Dense Graphical
Models [96.1052289276254]
This work introduces a new MAP-solver, based on the popular Dual Block-Coordinate Ascent principle.
Surprisingly, by making a small change to the low-performing solver, we derive the new solver MPLP++ that significantly outperforms all existing solvers by a large margin.
arXiv Detail & Related papers (2020-04-16T16:20:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.