An NMF-Based Building Block for Interpretable Neural Networks With
Continual Learning
- URL: http://arxiv.org/abs/2311.11485v1
- Date: Mon, 20 Nov 2023 02:00:33 GMT
- Title: An NMF-Based Building Block for Interpretable Neural Networks With
Continual Learning
- Authors: Brian K. Vogel
- Abstract summary: Existing learning methods often struggle to balance interpretability and predictive performance.
Our approach aims to strike a better balance between these two aspects through the use of a building block based on NMF.
- Score: 0.8158530638728501
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Existing learning methods often struggle to balance interpretability and
predictive performance. While models like nearest neighbors and non-negative
matrix factorization (NMF) offer high interpretability, their predictive
performance on supervised learning tasks is often limited. In contrast, neural
networks based on the multi-layer perceptron (MLP) support the modular
construction of expressive architectures and tend to have better recognition
accuracy but are often regarded as black boxes in terms of interpretability.
Our approach aims to strike a better balance between these two aspects through
the use of a building block based on NMF that incorporates supervised neural
network training methods to achieve high predictive performance while retaining
the desirable interpretability properties of NMF. We evaluate our Predictive
Factorized Coupling (PFC) block on small datasets and show that it achieves
competitive predictive performance with MLPs while also offering improved
interpretability. We demonstrate the benefits of this approach in various
scenarios, such as continual learning, training on non-i.i.d. data, and
knowledge removal after training. Additionally, we show examples of using the
PFC block to build more expressive architectures, including a fully-connected
residual network as well as a factorized recurrent neural network (RNN) that
performs competitively with vanilla RNNs while providing improved
interpretability. The PFC block uses an iterative inference algorithm that
converges to a fixed point, making it possible to trade off accuracy vs
computation after training but also currently preventing its use as a general
MLP replacement in some scenarios such as training on very large datasets. We
provide source code at https://github.com/bkvogel/pfc
Related papers
- Neural Reasoning Networks: Efficient Interpretable Neural Networks With Automatic Textual Explanations [45.974930902038494]
We propose a novel neuro-symbolic architecture, Neural Reasoning Networks (NRN), that is scalable and generates logically textual explanations for its predictions.
A training algorithm (R-NRN) learns the weights of the network as usual using descent optimization with backprop, but also learns the network structure itself using a bandit-based optimization.
R-NRN explanations are shorter than the compared approaches while producing more accurate feature importance scores.
arXiv Detail & Related papers (2024-10-10T14:27:12Z) - Convolutional Channel-wise Competitive Learning for the Forward-Forward
Algorithm [5.1246638322893245]
Forward-Forward (FF) algorithm has been proposed to alleviate the issues of backpropagation (BP) commonly used to train deep neural networks.
We take the main ideas of FF and improve them by leveraging channel-wise competitive learning in the context of convolutional neural networks for image classification tasks.
Our method outperforms recent FF-based models on image classification tasks, achieving testing errors of 0.58%, 7.69%, 21.89%, and 48.77% on MNIST, Fashion-MNIST, CIFAR-10 and CIFAR-100 respectively.
arXiv Detail & Related papers (2023-12-19T23:48:43Z) - Decouple Graph Neural Networks: Train Multiple Simple GNNs Simultaneously Instead of One [60.5818387068983]
Graph neural networks (GNN) suffer from severe inefficiency.
We propose to decouple a multi-layer GNN as multiple simple modules for more efficient training.
We show that the proposed framework is highly efficient with reasonable performance.
arXiv Detail & Related papers (2023-04-20T07:21:32Z) - Quantization-aware Interval Bound Propagation for Training Certifiably
Robust Quantized Neural Networks [58.195261590442406]
We study the problem of training and certifying adversarially robust quantized neural networks (QNNs)
Recent work has shown that floating-point neural networks that have been verified to be robust can become vulnerable to adversarial attacks after quantization.
We present quantization-aware interval bound propagation (QA-IBP), a novel method for training robust QNNs.
arXiv Detail & Related papers (2022-11-29T13:32:38Z) - Intelligence Processing Units Accelerate Neuromorphic Learning [52.952192990802345]
Spiking neural networks (SNNs) have achieved orders of magnitude improvement in terms of energy consumption and latency.
We present an IPU-optimized release of our custom SNN Python package, snnTorch.
arXiv Detail & Related papers (2022-11-19T15:44:08Z) - Comparative Analysis of Interval Reachability for Robust Implicit and
Feedforward Neural Networks [64.23331120621118]
We use interval reachability analysis to obtain robustness guarantees for implicit neural networks (INNs)
INNs are a class of implicit learning models that use implicit equations as layers.
We show that our approach performs at least as well as, and generally better than, applying state-of-the-art interval bound propagation methods to INNs.
arXiv Detail & Related papers (2022-04-01T03:31:27Z) - Dynamic Inference with Neural Interpreters [72.90231306252007]
We present Neural Interpreters, an architecture that factorizes inference in a self-attention network as a system of modules.
inputs to the model are routed through a sequence of functions in a way that is end-to-end learned.
We show that Neural Interpreters perform on par with the vision transformer using fewer parameters, while being transferrable to a new task in a sample efficient manner.
arXiv Detail & Related papers (2021-10-12T23:22:45Z) - FF-NSL: Feed-Forward Neural-Symbolic Learner [70.978007919101]
This paper introduces a neural-symbolic learning framework, called Feed-Forward Neural-Symbolic Learner (FF-NSL)
FF-NSL integrates state-of-the-art ILP systems based on the Answer Set semantics, with neural networks, in order to learn interpretable hypotheses from labelled unstructured data.
arXiv Detail & Related papers (2021-06-24T15:38:34Z) - Encoding the latent posterior of Bayesian Neural Networks for
uncertainty quantification [10.727102755903616]
We aim for efficient deep BNNs amenable to complex computer vision architectures.
We achieve this by leveraging variational autoencoders (VAEs) to learn the interaction and the latent distribution of the parameters at each network layer.
Our approach, Latent-Posterior BNN (LP-BNN), is compatible with the recent BatchEnsemble method, leading to highly efficient (in terms of computation and memory during both training and testing) ensembles.
arXiv Detail & Related papers (2020-12-04T19:50:09Z) - Effect of Architectures and Training Methods on the Performance of
Learned Video Frame Prediction [10.404162481860634]
Experimental results show that the residual FCNN architecture performs the best in terms of peak signal to noise ratio (PSNR) at the expense of higher training and test (inference) computational complexity.
The CRNN can be trained stably and very efficiently using the stateful truncated backpropagation through time procedure.
arXiv Detail & Related papers (2020-08-13T20:45:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.