An Order-Invariant and Interpretable Hierarchical Dilated Convolution
Neural Network for Chemical Fault Detection and Diagnosis
- URL: http://arxiv.org/abs/2302.06243v1
- Date: Mon, 13 Feb 2023 10:28:41 GMT
- Title: An Order-Invariant and Interpretable Hierarchical Dilated Convolution
Neural Network for Chemical Fault Detection and Diagnosis
- Authors: Mengxuan Li, Peng Peng, Min Wang, Hongwei Wang
- Abstract summary: Convolution neural network (CNN) is a popular deep learning algorithm with many successful applications in chemical fault detection and diagnosis tasks.
In this paper, we propose an order-invariant and interpretable hierarchical dilated convolution neural network (HDLCNN)
The proposed method provides interpretability by including the SHAP values to quantify feature contribution.
- Score: 7.226239130399725
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Fault detection and diagnosis is significant for reducing maintenance costs
and improving health and safety in chemical processes. Convolution neural
network (CNN) is a popular deep learning algorithm with many successful
applications in chemical fault detection and diagnosis tasks. However,
convolution layers in CNN are very sensitive to the order of features, which
can lead to instability in the processing of tabular data. Optimal order of
features result in better performance of CNN models but it is expensive to seek
such optimal order. In addition, because of the encapsulation mechanism of
feature extraction, most CNN models are opaque and have poor interpretability,
thus failing to identify root-cause features without human supervision. These
difficulties inevitably limit the performance and credibility of CNN methods.
In this paper, we propose an order-invariant and interpretable hierarchical
dilated convolution neural network (HDLCNN), which is composed by feature
clustering, dilated convolution and the shapley additive explanations (SHAP)
method. The novelty of HDLCNN lies in its capability of processing tabular data
with features of arbitrary order without seeking the optimal order, due to the
ability to agglomerate correlated features of feature clustering and the large
receptive field of dilated convolution. Then, the proposed method provides
interpretability by including the SHAP values to quantify feature contribution.
Therefore, the root-cause features can be identified as the features with the
highest contribution. Computational experiments are conducted on the Tennessee
Eastman chemical process benchmark dataset. Compared with the other methods,
the proposed HDLCNN-SHAP method achieves better performance on processing
tabular data with features of arbitrary order, detecting faults, and
identifying the root-cause features.
Related papers
- Learning local discrete features in explainable-by-design convolutional neural networks [0.0]
We introduce an explainable-by-design convolutional neural network (CNN) based on the lateral inhibition mechanism.
The model consists of the predictor, that is a high-accuracy CNN with residual or dense skip connections.
By collecting observations and directly calculating probabilities, we can explain causal relationships between motifs of adjacent levels.
arXiv Detail & Related papers (2024-10-31T18:39:41Z) - Hyperspectral Image Classification Based on Faster Residual Multi-branch Spiking Neural Network [6.166929138912052]
This paper builds a spiking neural network (SNN) based on the leaky integrate-and-fire (LIF) neuron model for HSI classification tasks.
SNN-SWMR requires a time step reduction of about 84%, training time, and testing time reduction of about 63% and 70% at the same accuracy.
arXiv Detail & Related papers (2024-09-18T00:51:01Z) - Unveiling the Power of Sparse Neural Networks for Feature Selection [60.50319755984697]
Sparse Neural Networks (SNNs) have emerged as powerful tools for efficient feature selection.
We show that SNNs trained with dynamic sparse training (DST) algorithms can achieve, on average, more than $50%$ memory and $55%$ FLOPs reduction.
Our findings show that feature selection with SNNs trained with DST algorithms can achieve, on average, more than $50%$ memory and $55%$ FLOPs reduction.
arXiv Detail & Related papers (2024-08-08T16:48:33Z) - Benign Overfitting in Deep Neural Networks under Lazy Training [72.28294823115502]
We show that when the data distribution is well-separated, DNNs can achieve Bayes-optimal test error for classification.
Our results indicate that interpolating with smoother functions leads to better generalization.
arXiv Detail & Related papers (2023-05-30T19:37:44Z) - Promises and Pitfalls of the Linearized Laplace in Bayesian Optimization [73.80101701431103]
The linearized-Laplace approximation (LLA) has been shown to be effective and efficient in constructing Bayesian neural networks.
We study the usefulness of the LLA in Bayesian optimization and highlight its strong performance and flexibility.
arXiv Detail & Related papers (2023-04-17T14:23:43Z) - Do deep neural networks have an inbuilt Occam's razor? [1.1470070927586016]
We show that structured data combined with an intrinsic Occam's razor-like inductive bias towards simple functions counteracts the exponential growth of functions with complexity.
This analysis reveals that structured data, combined with an intrinsic Occam's razor-like inductive bias towards (Kolmogorov) simple functions that is strong enough to counteract the exponential growth of functions with complexity, is a key to the success of DNNs.
arXiv Detail & Related papers (2023-04-13T16:58:21Z) - Non-Gradient Manifold Neural Network [79.44066256794187]
Deep neural network (DNN) generally takes thousands of iterations to optimize via gradient descent.
We propose a novel manifold neural network based on non-gradient optimization.
arXiv Detail & Related papers (2021-06-15T06:39:13Z) - CDiNN -Convex Difference Neural Networks [0.8122270502556374]
Neural networks with ReLU activation function have been shown to be universal function approximators learn function mapping as non-smooth functions.
New neural network architecture called ICNNs learn the output as a convex input.
arXiv Detail & Related papers (2021-03-31T17:31:16Z) - Online Limited Memory Neural-Linear Bandits with Likelihood Matching [53.18698496031658]
We study neural-linear bandits for solving problems where both exploration and representation learning play an important role.
We propose a likelihood matching algorithm that is resilient to catastrophic forgetting and is completely online.
arXiv Detail & Related papers (2021-02-07T14:19:07Z) - Supervised Learning for Non-Sequential Data: A Canonical Polyadic
Decomposition Approach [85.12934750565971]
Efficient modelling of feature interactions underpins supervised learning for non-sequential tasks.
To alleviate this issue, it has been proposed to implicitly represent the model parameters as a tensor.
For enhanced expressiveness, we generalize the framework to allow feature mapping to arbitrarily high-dimensional feature vectors.
arXiv Detail & Related papers (2020-01-27T22:38:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.