TFN: An Interpretable Neural Network with Time-Frequency Transform
Embedded for Intelligent Fault Diagnosis
- URL: http://arxiv.org/abs/2209.01992v2
- Date: Mon, 19 Jun 2023 08:55:08 GMT
- Title: TFN: An Interpretable Neural Network with Time-Frequency Transform
Embedded for Intelligent Fault Diagnosis
- Authors: Qian Chen, Xingjian Dong, Guowei Tu, Dong Wang, Baoxuan Zhao and Zhike
Peng
- Abstract summary: Convolutional Neural Networks (CNNs) are widely used in fault diagnosis of mechanical systems.
We propose a novel interpretable neural network termed as Time-Frequency Network (TFN), where the physically meaningful time-frequency transform (TFT) method is embedded into the traditional convolutional layer as an adaptive preprocessing layer.
In this study, four typical TFT methods are considered to formulate the TFNs and their effectiveness and interpretability are proved through three mechanical fault diagnosis experiments.
- Score: 6.812133175214715
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Convolutional Neural Networks (CNNs) are widely used in fault diagnosis of
mechanical systems due to their powerful feature extraction and classification
capabilities. However, the CNN is a typical black-box model, and the mechanism
of CNN's decision-making are not clear, which limits its application in
high-reliability-required fault diagnosis scenarios. To tackle this issue, we
propose a novel interpretable neural network termed as Time-Frequency Network
(TFN), where the physically meaningful time-frequency transform (TFT) method is
embedded into the traditional convolutional layer as an adaptive preprocessing
layer. This preprocessing layer named as time-frequency convolutional (TFconv)
layer, is constrained by a well-designed kernel function to extract
fault-related time-frequency information. It not only improves the diagnostic
performance but also reveals the logical foundation of the CNN prediction in
the frequency domain. Different TFT methods correspond to different kernel
functions of the TFconv layer. In this study, four typical TFT methods are
considered to formulate the TFNs and their effectiveness and interpretability
are proved through three mechanical fault diagnosis experiments. Experimental
results also show that the proposed TFconv layer can be easily generalized to
other CNNs with different depths. The code of TFN is available on
https://github.com/ChenQian0618/TFN.
Related papers
- How neural networks learn to classify chaotic time series [77.34726150561087]
We study the inner workings of neural networks trained to classify regular-versus-chaotic time series.
We find that the relation between input periodicity and activation periodicity is key for the performance of LKCNN models.
arXiv Detail & Related papers (2023-06-04T08:53:27Z) - A Scalable Walsh-Hadamard Regularizer to Overcome the Low-degree
Spectral Bias of Neural Networks [79.28094304325116]
Despite the capacity of neural nets to learn arbitrary functions, models trained through gradient descent often exhibit a bias towards simpler'' functions.
We show how this spectral bias towards low-degree frequencies can in fact hurt the neural network's generalization on real-world datasets.
We propose a new scalable functional regularization scheme that aids the neural network to learn higher degree frequencies.
arXiv Detail & Related papers (2023-05-16T20:06:01Z) - Continuous time recurrent neural networks: overview and application to
forecasting blood glucose in the intensive care unit [56.801856519460465]
Continuous time autoregressive recurrent neural networks (CTRNNs) are a deep learning model that account for irregular observations.
We demonstrate the application of these models to probabilistic forecasting of blood glucose in a critical care setting.
arXiv Detail & Related papers (2023-04-14T09:39:06Z) - Benign Overfitting for Two-layer ReLU Convolutional Neural Networks [60.19739010031304]
We establish algorithm-dependent risk bounds for learning two-layer ReLU convolutional neural networks with label-flipping noise.
We show that, under mild conditions, the neural network trained by gradient descent can achieve near-zero training loss and Bayes optimal test risk.
arXiv Detail & Related papers (2023-03-07T18:59:38Z) - NAF: Neural Attenuation Fields for Sparse-View CBCT Reconstruction [79.13750275141139]
This paper proposes a novel and fast self-supervised solution for sparse-view CBCT reconstruction.
The desired attenuation coefficients are represented as a continuous function of 3D spatial coordinates, parameterized by a fully-connected deep neural network.
A learning-based encoder entailing hash coding is adopted to help the network capture high-frequency details.
arXiv Detail & Related papers (2022-09-29T04:06:00Z) - Go Beyond Multiple Instance Neural Networks: Deep-learning Models based
on Local Pattern Aggregation [0.0]
convolutional neural networks (CNNs) have brought breakthroughs in processing clinical electrocardiograms (ECGs) and speaker-independent speech.
In this paper, we propose local pattern aggregation-based deep-learning models to effectively deal with both problems.
The novel network structure, called LPANet, has cropping and aggregation operations embedded into it.
arXiv Detail & Related papers (2022-05-28T13:18:18Z) - Lost Vibration Test Data Recovery Using Convolutional Neural Network: A
Case Study [0.0]
This paper proposes a CNN algorithm for the Alamosa Canyon Bridge as a real structure.
Three different CNN models were considered to predict one and two malfunctioned sensors.
The accuracy of the model was increased by adding a convolutional layer.
arXiv Detail & Related papers (2022-04-11T23:24:03Z) - Frequency Gating: Improved Convolutional Neural Networks for Speech
Enhancement in the Time-Frequency Domain [37.722450363816144]
We introduce a method, which we call Frequency Gating, to compute multiplicative weights for the kernels of the CNN.
Experiments with an autoencoder neural network with skip connections show that both local and frequency-wise gating outperform the baseline.
A loss function based on the extended short-time objective intelligibility score (ESTOI) is introduced, which we show to outperform the standard mean squared error (MSE) loss function.
arXiv Detail & Related papers (2020-11-08T22:04:00Z) - Accurate and Efficient Intracranial Hemorrhage Detection and Subtype
Classification in 3D CT Scans with Convolutional and Long Short-Term Memory
Neural Networks [20.4701676109641]
We present our system for the RSNA Intracranial Hemorrhage Detection challenge.
The proposed system is based on a lightweight deep neural network architecture composed of a convolutional neural network (CNN)
We report a weighted mean log loss of 0.04989 on the final test set, which places us in the top 30 ranking (2%) from a total of 1345 participants.
arXiv Detail & Related papers (2020-08-01T17:28:25Z) - Automate Obstructive Sleep Apnea Diagnosis Using Convolutional Neural
Networks [4.882119124419393]
This paper presents a CNN architecture with 1D convolutional and FCN layers for classification.
The proposed 1D CNN model achieves excellent classification results without manually preprocesssing PSG signals.
arXiv Detail & Related papers (2020-06-13T15:35:18Z) - Approximation and Non-parametric Estimation of ResNet-type Convolutional
Neural Networks [52.972605601174955]
We show a ResNet-type CNN can attain the minimax optimal error rates in important function classes.
We derive approximation and estimation error rates of the aformentioned type of CNNs for the Barron and H"older classes.
arXiv Detail & Related papers (2019-03-24T19:42:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.