PICNN: A Pathway towards Interpretable Convolutional Neural Networks
- URL: http://arxiv.org/abs/2312.12068v1
- Date: Tue, 19 Dec 2023 11:36:03 GMT
- Title: PICNN: A Pathway towards Interpretable Convolutional Neural Networks
- Authors: Wengang Guo, Jiayi Yang, Huilin Yin, Qijun Chen, Wei Ye
- Abstract summary: We introduce a novel pathway to alleviate the entanglement between filters and image classes.
We use the Bernoulli sampling to generate the filter-cluster assignment matrix from a learnable filter-class correspondence matrix.
We evaluate the effectiveness of our method on ten widely used network architectures.
- Score: 12.31424771480963
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Convolutional Neural Networks (CNNs) have exhibited great performance in
discriminative feature learning for complex visual tasks. Besides
discrimination power, interpretability is another important yet under-explored
property for CNNs. One difficulty in the CNN interpretability is that filters
and image classes are entangled. In this paper, we introduce a novel pathway to
alleviate the entanglement between filters and image classes. The proposed
pathway groups the filters in a late conv-layer of CNN into class-specific
clusters. Clusters and classes are in a one-to-one relationship. Specifically,
we use the Bernoulli sampling to generate the filter-cluster assignment matrix
from a learnable filter-class correspondence matrix. To enable end-to-end
optimization, we develop a novel reparameterization trick for handling the
non-differentiable Bernoulli sampling. We evaluate the effectiveness of our
method on ten widely used network architectures (including nine CNNs and a ViT)
and five benchmark datasets. Experimental results have demonstrated that our
method PICNN (the combination of standard CNNs with our proposed pathway)
exhibits greater interpretability than standard CNNs while achieving higher or
comparable discrimination power.
Related papers
- Training Convolutional Neural Networks with the Forward-Forward
algorithm [1.74440662023704]
Forward Forward (FF) algorithm has up to now only been used in fully connected networks.
We show how the FF paradigm can be extended to CNNs.
Our FF-trained CNN, featuring a novel spatially-extended labeling technique, achieves a classification accuracy of 99.16% on the MNIST hand-written digits dataset.
arXiv Detail & Related papers (2023-12-22T18:56:35Z) - Demystifying CNNs for Images by Matched Filters [13.121514086503591]
convolution neural networks (CNN) have been revolutionising the way we approach and use intelligent machines in the Big Data era.
CNNs have been put under scrutiny owing to their textitblack-box nature, as well as the lack of theoretical support and physical meanings of their operation.
This paper attempts to demystify the operation of CNNs by employing the perspective of matched filtering.
arXiv Detail & Related papers (2022-10-16T12:39:17Z) - Classification of diffraction patterns using a convolutional neural
network in single particle imaging experiments performed at X-ray
free-electron lasers [53.65540150901678]
Single particle imaging (SPI) at X-ray free electron lasers (XFELs) is particularly well suited to determine the 3D structure of particles in their native environment.
For a successful reconstruction, diffraction patterns originating from a single hit must be isolated from a large number of acquired patterns.
We propose to formulate this task as an image classification problem and solve it using convolutional neural network (CNN) architectures.
arXiv Detail & Related papers (2021-12-16T17:03:14Z) - Graph Neural Networks with Adaptive Frequency Response Filter [55.626174910206046]
We develop a graph neural network framework AdaGNN with a well-smooth adaptive frequency response filter.
We empirically validate the effectiveness of the proposed framework on various benchmark datasets.
arXiv Detail & Related papers (2021-04-26T19:31:21Z) - The Mind's Eye: Visualizing Class-Agnostic Features of CNNs [92.39082696657874]
We propose an approach to visually interpret CNN features given a set of images by creating corresponding images that depict the most informative features of a specific layer.
Our method uses a dual-objective activation and distance loss, without requiring a generator network nor modifications to the original model.
arXiv Detail & Related papers (2021-01-29T07:46:39Z) - Convolutional Neural Networks for Multispectral Image Cloud Masking [7.812073412066698]
Convolutional neural networks (CNN) have proven to be state of the art methods for many image classification tasks.
We study the use of different CNN architectures for cloud masking of Proba-V multispectral images.
arXiv Detail & Related papers (2020-12-09T21:33:20Z) - Training Interpretable Convolutional Neural Networks by Differentiating
Class-specific Filters [64.46270549587004]
Convolutional neural networks (CNNs) have been successfully used in a range of tasks.
CNNs are often viewed as "black-box" and lack of interpretability.
We propose a novel strategy to train interpretable CNNs by encouraging class-specific filters.
arXiv Detail & Related papers (2020-07-16T09:12:26Z) - Curriculum By Smoothing [52.08553521577014]
Convolutional Neural Networks (CNNs) have shown impressive performance in computer vision tasks such as image classification, detection, and segmentation.
We propose an elegant curriculum based scheme that smoothes the feature embedding of a CNN using anti-aliasing or low-pass filters.
As the amount of information in the feature maps increases during training, the network is able to progressively learn better representations of the data.
arXiv Detail & Related papers (2020-03-03T07:27:44Z) - Approximation and Non-parametric Estimation of ResNet-type Convolutional
Neural Networks [52.972605601174955]
We show a ResNet-type CNN can attain the minimax optimal error rates in important function classes.
We derive approximation and estimation error rates of the aformentioned type of CNNs for the Barron and H"older classes.
arXiv Detail & Related papers (2019-03-24T19:42:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.