An Order-Invariant and Interpretable Hierarchical Dilated Convolution
Neural Network for Chemical Fault Detection and Diagnosis
- URL: http://arxiv.org/abs/2302.06243v1
- Date: Mon, 13 Feb 2023 10:28:41 GMT
- Title: An Order-Invariant and Interpretable Hierarchical Dilated Convolution
Neural Network for Chemical Fault Detection and Diagnosis
- Authors: Mengxuan Li, Peng Peng, Min Wang, Hongwei Wang
- Abstract summary: Convolution neural network (CNN) is a popular deep learning algorithm with many successful applications in chemical fault detection and diagnosis tasks.
In this paper, we propose an order-invariant and interpretable hierarchical dilated convolution neural network (HDLCNN)
The proposed method provides interpretability by including the SHAP values to quantify feature contribution.
- Score: 7.226239130399725
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Fault detection and diagnosis is significant for reducing maintenance costs
and improving health and safety in chemical processes. Convolution neural
network (CNN) is a popular deep learning algorithm with many successful
applications in chemical fault detection and diagnosis tasks. However,
convolution layers in CNN are very sensitive to the order of features, which
can lead to instability in the processing of tabular data. Optimal order of
features result in better performance of CNN models but it is expensive to seek
such optimal order. In addition, because of the encapsulation mechanism of
feature extraction, most CNN models are opaque and have poor interpretability,
thus failing to identify root-cause features without human supervision. These
difficulties inevitably limit the performance and credibility of CNN methods.
In this paper, we propose an order-invariant and interpretable hierarchical
dilated convolution neural network (HDLCNN), which is composed by feature
clustering, dilated convolution and the shapley additive explanations (SHAP)
method. The novelty of HDLCNN lies in its capability of processing tabular data
with features of arbitrary order without seeking the optimal order, due to the
ability to agglomerate correlated features of feature clustering and the large
receptive field of dilated convolution. Then, the proposed method provides
interpretability by including the SHAP values to quantify feature contribution.
Therefore, the root-cause features can be identified as the features with the
highest contribution. Computational experiments are conducted on the Tennessee
Eastman chemical process benchmark dataset. Compared with the other methods,
the proposed HDLCNN-SHAP method achieves better performance on processing
tabular data with features of arbitrary order, detecting faults, and
identifying the root-cause features.
Related papers
- Enhancing GNNs Performance on Combinatorial Optimization by Recurrent Feature Update [0.09986418756990156]
We introduce a novel algorithm, denoted hereafter as QRF-GNN, leveraging the power of GNNs to efficiently solve Combinatorial optimization (CO) problems.
It relies on unsupervised learning by minimizing the loss function derived from QUBO relaxation.
Results of experiments show that QRF-GNN drastically surpasses existing learning-based approaches and is comparable to the state-of-the-art conventionals.
arXiv Detail & Related papers (2024-07-23T13:34:35Z) - Benign Overfitting in Deep Neural Networks under Lazy Training [72.28294823115502]
We show that when the data distribution is well-separated, DNNs can achieve Bayes-optimal test error for classification.
Our results indicate that interpolating with smoother functions leads to better generalization.
arXiv Detail & Related papers (2023-05-30T19:37:44Z) - Promises and Pitfalls of the Linearized Laplace in Bayesian Optimization [73.80101701431103]
The linearized-Laplace approximation (LLA) has been shown to be effective and efficient in constructing Bayesian neural networks.
We study the usefulness of the LLA in Bayesian optimization and highlight its strong performance and flexibility.
arXiv Detail & Related papers (2023-04-17T14:23:43Z) - Do deep neural networks have an inbuilt Occam's razor? [1.1470070927586016]
We show that structured data combined with an intrinsic Occam's razor-like inductive bias towards simple functions counteracts the exponential growth of functions with complexity.
This analysis reveals that structured data, combined with an intrinsic Occam's razor-like inductive bias towards (Kolmogorov) simple functions that is strong enough to counteract the exponential growth of functions with complexity, is a key to the success of DNNs.
arXiv Detail & Related papers (2023-04-13T16:58:21Z) - SA-CNN: Application to text categorization issues using simulated
annealing-based convolutional neural network optimization [0.0]
Convolutional neural networks (CNNs) are a representative class of deep learning algorithms.
We introduce SA-CNN neural networks for text classification tasks based on Text-CNN neural networks.
arXiv Detail & Related papers (2023-03-13T14:27:34Z) - Accurate, reliable and interpretable solubility prediction of druglike
molecules with attention pooling and Bayesian learning [1.8275108630751844]
In silico prediction of solubility has been studied for its utility in virtual screening and lead optimization.
Recently, machine learning (ML) methods using experimental data has been popular because physics-based methods are not suitable for high- throughput tasks.
In this paper, we develop graph neural networks (GNNs) with the self-attention readout layer to improve prediction performance.
arXiv Detail & Related papers (2022-09-29T07:48:10Z) - Training Feedback Spiking Neural Networks by Implicit Differentiation on
the Equilibrium State [66.2457134675891]
Spiking neural networks (SNNs) are brain-inspired models that enable energy-efficient implementation on neuromorphic hardware.
Most existing methods imitate the backpropagation framework and feedforward architectures for artificial neural networks.
We propose a novel training method that does not rely on the exact reverse of the forward computation.
arXiv Detail & Related papers (2021-09-29T07:46:54Z) - Non-Gradient Manifold Neural Network [79.44066256794187]
Deep neural network (DNN) generally takes thousands of iterations to optimize via gradient descent.
We propose a novel manifold neural network based on non-gradient optimization.
arXiv Detail & Related papers (2021-06-15T06:39:13Z) - CDiNN -Convex Difference Neural Networks [0.8122270502556374]
Neural networks with ReLU activation function have been shown to be universal function approximators learn function mapping as non-smooth functions.
New neural network architecture called ICNNs learn the output as a convex input.
arXiv Detail & Related papers (2021-03-31T17:31:16Z) - Online Limited Memory Neural-Linear Bandits with Likelihood Matching [53.18698496031658]
We study neural-linear bandits for solving problems where both exploration and representation learning play an important role.
We propose a likelihood matching algorithm that is resilient to catastrophic forgetting and is completely online.
arXiv Detail & Related papers (2021-02-07T14:19:07Z) - Supervised Learning for Non-Sequential Data: A Canonical Polyadic
Decomposition Approach [85.12934750565971]
Efficient modelling of feature interactions underpins supervised learning for non-sequential tasks.
To alleviate this issue, it has been proposed to implicitly represent the model parameters as a tensor.
For enhanced expressiveness, we generalize the framework to allow feature mapping to arbitrarily high-dimensional feature vectors.
arXiv Detail & Related papers (2020-01-27T22:38:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.