Explaining Neural Network Predictions for Functional Data Using
Principal Component Analysis and Feature Importance
- URL: http://arxiv.org/abs/2010.12063v1
- Date: Thu, 15 Oct 2020 22:33:21 GMT
- Title: Explaining Neural Network Predictions for Functional Data Using
Principal Component Analysis and Feature Importance
- Authors: Katherine Goode, Daniel Ries, Joshua Zollweg
- Abstract summary: We propose a procedure for explaining machine learning models fit using functional data.
We demonstrate the technique by explaining neural networks fit to explosion optical spectral-temporal signatures.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Optical spectral-temporal signatures extracted from videos of explosions
provide information for identifying characteristics of the corresponding
explosive devices. Currently, the identification is done using heuristic
algorithms and direct subject matter expert review. An improvement in
predictive performance may be obtained by using machine learning, but this
application lends itself to high consequence national security decisions, so it
is not only important to provide high accuracy but clear explanations for the
predictions to garner confidence in the model. While much work has been done to
develop explainability methods for machine learning models, not much of the
work focuses on situations with input variables of the form of functional data
such optical spectral-temporal signatures. We propose a procedure for
explaining machine learning models fit using functional data that accounts for
the functional nature the data. Our approach makes use of functional principal
component analysis (fPCA) and permutation feature importance (PFI). fPCA is
used to transform the functions to create uncorrelated functional principal
components (fPCs). The model is trained using the fPCs as inputs, and PFI is
applied to identify the fPCs important to the model for prediction.
Visualizations are used to interpret the variability explained by the fPCs that
are found to be important by PFI to determine the aspects of the functions that
are important for prediction. We demonstrate the technique by explaining neural
networks fit to explosion optical spectral-temporal signatures for predicting
characteristics of the explosive devices.
Related papers
- An Explainable Pipeline for Machine Learning with Functional Data [0.1874930567916036]
We consider two applications from high-consequence spaces with objectives of making predictions using functional data inputs.
One application aims to classify material types to identify explosive materials given hyperspectral computed tomography scans of the materials.
The other application considers the forensics science task of connecting an inkjet printed document to the source printer using color signatures extracted by Raman spectroscopy.
arXiv Detail & Related papers (2025-01-10T23:33:15Z) - Localized Gaussians as Self-Attention Weights for Point Clouds Correspondence [92.07601770031236]
We investigate semantically meaningful patterns in the attention heads of an encoder-only Transformer architecture.
We find that fixing the attention weights not only accelerates the training process but also enhances the stability of the optimization.
arXiv Detail & Related papers (2024-09-20T07:41:47Z) - Demystifying Functional Random Forests: Novel Explainability Tools for Model Transparency in High-Dimensional Spaces [0.0]
This paper introduces a novel suite of explainability tools to illuminate the inner mechanisms of Functional Random Forests (FRF)
These tools collectively enhance the transparency of FRF models by providing a detailed analysis of how individual FPCs contribute to model predictions.
arXiv Detail & Related papers (2024-08-22T10:52:32Z) - A Mechanistic Interpretation of Arithmetic Reasoning in Language Models
using Causal Mediation Analysis [128.0532113800092]
We present a mechanistic interpretation of Transformer-based LMs on arithmetic questions.
This provides insights into how information related to arithmetic is processed by LMs.
arXiv Detail & Related papers (2023-05-24T11:43:47Z) - How Spurious Features Are Memorized: Precise Analysis for Random and NTK Features [19.261178173399784]
We consider spurious features that are uncorrelated with the learning task.
We provide a precise characterization of how they are memorized via two separate terms.
We prove that the memorization of spurious features weakens as the generalization capability increases.
arXiv Detail & Related papers (2023-05-20T05:27:41Z) - Information-Theoretic Odometry Learning [83.36195426897768]
We propose a unified information theoretic framework for learning-motivated methods aimed at odometry estimation.
The proposed framework provides an elegant tool for performance evaluation and understanding in information-theoretic language.
arXiv Detail & Related papers (2022-03-11T02:37:35Z) - SparseBERT: Rethinking the Importance Analysis in Self-attention [107.68072039537311]
Transformer-based models are popular for natural language processing (NLP) tasks due to its powerful capacity.
Attention map visualization of a pre-trained model is one direct method for understanding self-attention mechanism.
We propose a Differentiable Attention Mask (DAM) algorithm, which can be also applied in guidance of SparseBERT design.
arXiv Detail & Related papers (2021-02-25T14:13:44Z) - Estimating Structural Target Functions using Machine Learning and
Influence Functions [103.47897241856603]
We propose a new framework for statistical machine learning of target functions arising as identifiable functionals from statistical models.
This framework is problem- and model-agnostic and can be used to estimate a broad variety of target parameters of interest in applied statistics.
We put particular focus on so-called coarsening at random/doubly robust problems with partially unobserved information.
arXiv Detail & Related papers (2020-08-14T16:48:29Z) - Controlling for sparsity in sparse factor analysis models: adaptive
latent feature sharing for piecewise linear dimensionality reduction [2.896192909215469]
We propose a simple and tractable parametric feature allocation model which can address key limitations of current latent feature decomposition techniques.
We derive a novel adaptive Factor analysis (aFA), as well as, an adaptive probabilistic principle component analysis (aPPCA) capable of flexible structure discovery and dimensionality reduction.
We show that aPPCA and aFA can infer interpretable high level features both when applied on raw MNIST and when applied for interpreting autoencoder features.
arXiv Detail & Related papers (2020-06-22T16:09:11Z) - Explaining Black Box Predictions and Unveiling Data Artifacts through
Influence Functions [55.660255727031725]
Influence functions explain the decisions of a model by identifying influential training examples.
We conduct a comparison between influence functions and common word-saliency methods on representative tasks.
We develop a new measure based on influence functions that can reveal artifacts in training data.
arXiv Detail & Related papers (2020-05-14T00:45:23Z) - On the Estimation of Complex Circuits Functional Failure Rate by Machine
Learning Techniques [0.16311150636417257]
De-Rating or Vulnerability Factors are a major feature of failure analysis efforts mandated by today's Functional Safety requirements.
New approach is proposed which uses Machine Learning to estimate the Functional De-Rating of individual flip-flops.
arXiv Detail & Related papers (2020-02-18T15:18:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.