Using explainability to design physics-aware CNNs for solving subsurface
inverse problems
- URL: http://arxiv.org/abs/2211.08651v2
- Date: Fri, 31 Mar 2023 22:50:01 GMT
- Title: Using explainability to design physics-aware CNNs for solving subsurface
inverse problems
- Authors: Jodie Crocker (1), Krishna Kumar (1), Brady R. Cox (2) ((1) The
University of Texas at Austin, (2) Utah State University)
- Abstract summary: We present a novel method of using explainability techniques to design physics-aware neural networks.
We demonstrate our approach by developing a convolutional neural network (CNN) for solving an inverse problem for shallow subsurface imaging.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present a novel method of using explainability techniques to design
physics-aware neural networks. We demonstrate our approach by developing a
convolutional neural network (CNN) for solving an inverse problem for shallow
subsurface imaging. Although CNNs have gained popularity in recent years across
many fields, the development of CNNs remains an art, as there are no clear
guidelines regarding the selection of hyperparameters that will yield the best
network. While optimization algorithms may be used to select hyperparameters
automatically, these methods focus on developing networks with high predictive
accuracy while disregarding model explainability (descriptive accuracy).
However, the field of Explainable Artificial Intelligence (XAI) addresses the
absence of model explainability by providing tools that allow developers to
evaluate the internal logic of neural networks. In this study, we use the
explainability methods Score-CAM and Deep SHAP to select hyperparameters, such
as kernel sizes and network depth, to develop a physics-aware CNN for shallow
subsurface imaging. We begin with a relatively deep Encoder-Decoder network,
which uses surface wave dispersion images as inputs and generates 2D shear wave
velocity subsurface images as outputs. Through model explanations, we
ultimately find that a shallow CNN using two convolutional layers with an
atypical kernel size of 3x1 yields comparable predictive accuracy but with
increased descriptive accuracy. We also show that explainability methods can be
used to evaluate the network's complexity and decision-making. We believe this
method can be used to develop neural networks with high predictive accuracy
while also providing inherent explainability.
Related papers
- An Analysis Framework for Understanding Deep Neural Networks Based on Network Dynamics [11.44947569206928]
Deep neural networks (DNNs) maximize information extraction by rationally allocating the proportion of neurons in different modes across deep layers.
This framework provides a unified explanation for fundamental DNN behaviors such as the "flat minima effect," "grokking," and double descent phenomena.
arXiv Detail & Related papers (2025-01-05T04:23:21Z) - PiLocNet: Physics-informed neural network on 3D localization with rotating point spread function [3.029152208453665]
We propose a novel enhancement of our previously introduced localization neural network, LocNet.
The improved network is a physics-informed neural network (PINN) that we call PiLocNet.
Although the paper focuses on the use of single-lobe rotating PSF to encode the full 3D source location, we expect the method to be widely applicable to other PSFs and imaging problems.
arXiv Detail & Related papers (2024-10-17T07:49:23Z) - GINN-KAN: Interpretability pipelining with applications in Physics Informed Neural Networks [5.2969467015867915]
We introduce the concept of interpretability pipelineing, to incorporate multiple interpretability techniques to outperform each individual technique.
We evaluate two recent models selected for their potential to incorporate interpretability into standard neural network architectures.
We introduce a novel interpretable neural network GINN-KAN that synthesizes the advantages of both models.
arXiv Detail & Related papers (2024-08-27T04:57:53Z) - Graph Neural Networks for Learning Equivariant Representations of Neural Networks [55.04145324152541]
We propose to represent neural networks as computational graphs of parameters.
Our approach enables a single model to encode neural computational graphs with diverse architectures.
We showcase the effectiveness of our method on a wide range of tasks, including classification and editing of implicit neural representations.
arXiv Detail & Related papers (2024-03-18T18:01:01Z) - NAR-Former: Neural Architecture Representation Learning towards Holistic
Attributes Prediction [37.357949900603295]
We propose a neural architecture representation model that can be used to estimate attributes holistically.
Experiment results show that our proposed framework can be used to predict the latency and accuracy attributes of both cell architectures and whole deep neural networks.
arXiv Detail & Related papers (2022-11-15T10:15:21Z) - Parameter Convex Neural Networks [13.42851919291587]
We propose the exponential multilayer neural network (EMLP) which is convex with regard to the parameters of the neural network under some conditions.
For late experiments, we use the same architecture to make the exponential graph convolutional network (EGCN) and do the experiment on the graph classificaion dataset.
arXiv Detail & Related papers (2022-06-11T16:44:59Z) - Data-driven emergence of convolutional structure in neural networks [83.4920717252233]
We show how fully-connected neural networks solving a discrimination task can learn a convolutional structure directly from their inputs.
By carefully designing data models, we show that the emergence of this pattern is triggered by the non-Gaussian, higher-order local structure of the inputs.
arXiv Detail & Related papers (2022-02-01T17:11:13Z) - CondenseNeXt: An Ultra-Efficient Deep Neural Network for Embedded
Systems [0.0]
A Convolutional Neural Network (CNN) is a class of Deep Neural Network (DNN) widely used in the analysis of visual images captured by an image sensor.
In this paper, we propose a neoteric variant of deep convolutional neural network architecture to ameliorate the performance of existing CNN architectures for real-time inference on embedded systems.
arXiv Detail & Related papers (2021-12-01T18:20:52Z) - Spiking Neural Networks -- Part II: Detecting Spatio-Temporal Patterns [38.518936229794214]
Spiking Neural Networks (SNNs) have the unique ability to detect information in encoded-temporal signals.
We review models and training algorithms for the dominant approach that considers SNNs as a Recurrent Neural Network (RNN)
We describe an alternative approach that relies on probabilistic models for spiking neurons, allowing the derivation of local learning rules via gradient estimates.
arXiv Detail & Related papers (2020-10-27T11:47:42Z) - Modeling from Features: a Mean-field Framework for Over-parameterized
Deep Neural Networks [54.27962244835622]
This paper proposes a new mean-field framework for over- parameterized deep neural networks (DNNs)
In this framework, a DNN is represented by probability measures and functions over its features in the continuous limit.
We illustrate the framework via the standard DNN and the Residual Network (Res-Net) architectures.
arXiv Detail & Related papers (2020-07-03T01:37:16Z) - Rectified Linear Postsynaptic Potential Function for Backpropagation in
Deep Spiking Neural Networks [55.0627904986664]
Spiking Neural Networks (SNNs) usetemporal spike patterns to represent and transmit information, which is not only biologically realistic but also suitable for ultra-low-power event-driven neuromorphic implementation.
This paper investigates the contribution of spike timing dynamics to information encoding, synaptic plasticity and decision making, providing a new perspective to design of future DeepSNNs and neuromorphic hardware systems.
arXiv Detail & Related papers (2020-03-26T11:13:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.