Widely Linear Matched Filter: A Lynchpin towards the Interpretability of
Complex-valued CNNs
- URL: http://arxiv.org/abs/2401.16729v2
- Date: Wed, 31 Jan 2024 05:11:54 GMT
- Title: Widely Linear Matched Filter: A Lynchpin towards the Interpretability of
Complex-valued CNNs
- Authors: Qingchen Wang, Zhe Li, Zdenka Babic, Wei Deng, Ljubi\v{s}a
Stankovi\'c, Danilo P. Mandic
- Abstract summary: We introduce a general WLMF paradigm, provide its solution and undertake analysis of its performance.
For rigor, our WLMF solution is derived without imposing any assumption on the probability density of noise.
This serves to revisit the convolution-activation-pooling chain in complex-valued CNNs through the lens of matched filtering.
- Score: 19.291619185044173
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A recent study on the interpretability of real-valued convolutional neural
networks (CNNs) {Stankovic_Mandic_2023CNN} has revealed a direct and physically
meaningful link with the task of finding features in data through matched
filters. However, applying this paradigm to illuminate the interpretability of
complex-valued CNNs meets a formidable obstacle: the extension of matched
filtering to a general class of noncircular complex-valued data, referred to
here as the widely linear matched filter (WLMF), has been only implicit in the
literature. To this end, to establish the interpretability of the operation of
complex-valued CNNs, we introduce a general WLMF paradigm, provide its solution
and undertake analysis of its performance. For rigor, our WLMF solution is
derived without imposing any assumption on the probability density of noise.
The theoretical advantages of the WLMF over its standard strictly linear
counterpart (SLMF) are provided in terms of their output signal-to-noise-ratios
(SNRs), with WLMF consistently exhibiting enhanced SNR. Moreover, the lower
bound on the SNR gain of WLMF is derived, together with condition to attain
this bound. This serves to revisit the convolution-activation-pooling chain in
complex-valued CNNs through the lens of matched filtering, which reveals the
potential of WLMFs to provide physical interpretability and enhance
explainability of general complex-valued CNNs. Simulations demonstrate the
agreement between the theoretical and numerical results.
Related papers
- Inferring Dynamic Networks from Marginals with Iterative Proportional Fitting [57.487936697747024]
A common network inference problem, arising from real-world data constraints, is how to infer a dynamic network from its time-aggregated adjacency matrix.
We introduce a principled algorithm that guarantees IPF converges under minimal changes to the network structure.
arXiv Detail & Related papers (2024-02-28T20:24:56Z) - PICNN: A Pathway towards Interpretable Convolutional Neural Networks [12.31424771480963]
We introduce a novel pathway to alleviate the entanglement between filters and image classes.
We use the Bernoulli sampling to generate the filter-cluster assignment matrix from a learnable filter-class correspondence matrix.
We evaluate the effectiveness of our method on ten widely used network architectures.
arXiv Detail & Related papers (2023-12-19T11:36:03Z) - Benign Overfitting in Deep Neural Networks under Lazy Training [72.28294823115502]
We show that when the data distribution is well-separated, DNNs can achieve Bayes-optimal test error for classification.
Our results indicate that interpolating with smoother functions leads to better generalization.
arXiv Detail & Related papers (2023-05-30T19:37:44Z) - SymNMF-Net for The Symmetric NMF Problem [62.44067422984995]
We propose a neural network called SymNMF-Net for the Symmetric NMF problem.
We show that the inference of each block corresponds to a single iteration of the optimization.
Empirical results on real-world datasets demonstrate the superiority of our SymNMF-Net.
arXiv Detail & Related papers (2022-05-26T08:17:39Z) - Comparative Analysis of Interval Reachability for Robust Implicit and
Feedforward Neural Networks [64.23331120621118]
We use interval reachability analysis to obtain robustness guarantees for implicit neural networks (INNs)
INNs are a class of implicit learning models that use implicit equations as layers.
We show that our approach performs at least as well as, and generally better than, applying state-of-the-art interval bound propagation methods to INNs.
arXiv Detail & Related papers (2022-04-01T03:31:27Z) - Batch Normalization Tells You Which Filter is Important [49.903610684578716]
We propose a simple yet effective filter pruning method by evaluating the importance of each filter based on the BN parameters of pre-trained CNNs.
The experimental results on CIFAR-10 and ImageNet demonstrate that the proposed method can achieve outstanding performance.
arXiv Detail & Related papers (2021-12-02T12:04:59Z) - Self-interpretable Convolutional Neural Networks for Text Classification [5.55878488884108]
This paper develops an approach for interpreting convolutional neural networks for text classification problems by exploiting the local-linear models inherent in ReLU-DNNs.
We show that our proposed technique produce parsimonious models that are self-interpretable and have comparable performance with respect to a more complex CNN model.
arXiv Detail & Related papers (2021-05-18T15:19:59Z) - Generalized Approach to Matched Filtering using Neural Networks [4.535489275919893]
We make a key observation on the relationship between the emerging deep learning and the traditional techniques.
matched filtering is formally equivalent to a particular neural network.
We show that the proposed neural network architecture can outperform matched filtering.
arXiv Detail & Related papers (2021-04-08T17:59:07Z) - A machine learning framework for LES closure terms [0.0]
We derive a consistent framework for LES closure models, with special emphasis laid upon the incorporation of implicit discretization-based filters and numerical approximation errors.
We compute the exact closure terms for the different LES filter functions from direct numerical simulation results of decaying homogeneous isotropic turbulence.
For the given application, the GRU architecture clearly outperforms the networks in terms of accuracy.
arXiv Detail & Related papers (2020-10-01T08:42:37Z) - Learning Likelihoods with Conditional Normalizing Flows [54.60456010771409]
Conditional normalizing flows (CNFs) are efficient in sampling and inference.
We present a study of CNFs where the base density to output space mapping is conditioned on an input x, to model conditional densities p(y|x)
arXiv Detail & Related papers (2019-11-29T19:17:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.