Understanding of Kernels in CNN Models by Suppressing Irrelevant Visual
Features in Images
- URL: http://arxiv.org/abs/2108.11054v1
- Date: Wed, 25 Aug 2021 05:48:44 GMT
- Title: Understanding of Kernels in CNN Models by Suppressing Irrelevant Visual
Features in Images
- Authors: Jia-Xin Zhuang, Wanying Tao, Jianfei Xing, Wei Shi, Ruixuan Wang,
Wei-shi Zheng
- Abstract summary: The lack of precisely interpreting kernels in convolutional neural networks (CNNs) is one main obstacle to wide applications of deep learning models in real scenarios.
A simple yet effective optimization method is proposed to interpret the activation of any kernel of interest in CNN models.
- Score: 55.60727570036073
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning models have shown their superior performance in various vision
tasks. However, the lack of precisely interpreting kernels in convolutional
neural networks (CNNs) is becoming one main obstacle to wide applications of
deep learning models in real scenarios. Although existing interpretation
methods may find certain visual patterns which are associated with the
activation of a specific kernel, those visual patterns may not be specific or
comprehensive enough for interpretation of a specific activation of kernel of
interest. In this paper, a simple yet effective optimization method is proposed
to interpret the activation of any kernel of interest in CNN models. The basic
idea is to simultaneously preserve the activation of the specific kernel and
suppress the activation of all other kernels at the same layer. In this way,
only visual information relevant to the activation of the specific kernel is
remained in the input. Consistent visual information from multiple modified
inputs would help users understand what kind of features are specifically
associated with specific kernel. Comprehensive evaluation shows that the
proposed method can help better interpret activation of specific kernels than
widely used methods, even when two kernels have very similar activation regions
from the same input image.
Related papers
- Operator theory, kernels, and Feedforward Neural Networks [0.0]
We show how specific families of positive definite kernels serve as powerful tools in analyses of algorithms for multiple layer feedforward Neural Network models.
Our focus is on particular kernels that adapt well to learning algorithms for data-sets/features which display intrinsic self-similarities at feedforward iteration of scaling.
arXiv Detail & Related papers (2023-01-03T19:30:31Z) - Joint Embedding Self-Supervised Learning in the Kernel Regime [21.80241600638596]
Self-supervised learning (SSL) produces useful representations of data without access to any labels for classifying the data.
We extend this framework to incorporate algorithms based on kernel methods where embeddings are constructed by linear maps acting on the feature space of a kernel.
We analyze our kernel model on small datasets to identify common features of self-supervised learning algorithms and gain theoretical insights into their performance on downstream tasks.
arXiv Detail & Related papers (2022-09-29T15:53:19Z) - Inducing Gaussian Process Networks [80.40892394020797]
We propose inducing Gaussian process networks (IGN), a simple framework for simultaneously learning the feature space as well as the inducing points.
The inducing points, in particular, are learned directly in the feature space, enabling a seamless representation of complex structured domains.
We report on experimental results for real-world data sets showing that IGNs provide significant advances over state-of-the-art methods.
arXiv Detail & Related papers (2022-04-21T05:27:09Z) - Kernel Continual Learning [117.79080100313722]
kernel continual learning is a simple but effective variant of continual learning to tackle catastrophic forgetting.
episodic memory unit stores a subset of samples for each task to learn task-specific classifiers based on kernel ridge regression.
variational random features to learn a data-driven kernel for each task.
arXiv Detail & Related papers (2021-07-12T22:09:30Z) - Random Features for the Neural Tangent Kernel [57.132634274795066]
We propose an efficient feature map construction of the Neural Tangent Kernel (NTK) of fully-connected ReLU network.
We show that dimension of the resulting features is much smaller than other baseline feature map constructions to achieve comparable error bounds both in theory and practice.
arXiv Detail & Related papers (2021-04-03T09:08:12Z) - Neural Generalization of Multiple Kernel Learning [2.064612766965483]
Multiple Kernel Learning is a conventional way to learn the kernel function in kernel-based methods.
Deep learning models can learn complex functions by applying nonlinear transformations to data through several layers.
We show that a typical MKL algorithm can be interpreted as a one-layer neural network with linear activation functions.
arXiv Detail & Related papers (2021-02-26T07:28:37Z) - On Approximation in Deep Convolutional Networks: a Kernel Perspective [12.284934135116515]
We study the success of deep convolutional networks on tasks involving high-dimensional data such as images or audio.
We study this theoretically and empirically through the lens of kernel methods, by considering multi-layer convolutional kernels.
We find that while expressive kernels operating on input patches are important at the first layer, simpler kernels can suffice in higher layers for good performance.
arXiv Detail & Related papers (2021-02-19T17:03:42Z) - Bayesian Sparse Factor Analysis with Kernelized Observations [67.60224656603823]
Multi-view problems can be faced with latent variable models.
High-dimensionality and non-linear issues are traditionally handled by kernel methods.
We propose merging both approaches into single model.
arXiv Detail & Related papers (2020-06-01T14:25:38Z) - Avoiding Kernel Fixed Points: Computing with ELU and GELU Infinite
Networks [12.692279981822011]
We derive the covariance functions of multi-layer perceptrons with exponential linear units (ELU) and Gaussian error linear units (GELU)
We analyse the fixed-point dynamics of iterated kernels corresponding to a broad range of activation functions.
We find that unlike some previously studied neural network kernels, these new kernels exhibit non-trivial fixed-point dynamics.
arXiv Detail & Related papers (2020-02-20T01:25:39Z) - Learning Class Regularized Features for Action Recognition [68.90994813947405]
We introduce a novel method named Class Regularization that performs class-based regularization of layer activations.
We show that using Class Regularization blocks in state-of-the-art CNN architectures for action recognition leads to systematic improvement gains of 1.8%, 1.2% and 1.4% on the Kinetics, UCF-101 and HMDB-51 datasets, respectively.
arXiv Detail & Related papers (2020-02-07T07:27:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.