Convolutional Neural Network with Convolutional Block Attention Module
for Finger Vein Recognition
- URL: http://arxiv.org/abs/2202.06673v1
- Date: Mon, 14 Feb 2022 12:59:23 GMT
- Title: Convolutional Neural Network with Convolutional Block Attention Module
for Finger Vein Recognition
- Authors: Zhongxia Zhang and Mingwen Wang
- Abstract summary: We propose a lightweight convolutional neural network with a convolutional block attention module (CBAM) for finger vein recognition.
The experiments are carried out on two publicly available databases and the results demonstrate that the proposed method achieves a stable, highly accurate, and robust performance in multimodal finger recognition.
- Score: 4.035753155957698
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Convolutional neural networks have become a popular research in the field of
finger vein recognition because of their powerful image feature representation.
However, most researchers focus on improving the performance of the network by
increasing the CNN depth and width, which often requires high computational
effort. Moreover, we can notice that not only the importance of pixels in
different channels is different, but also the importance of pixels in different
positions of the same channel is different. To reduce the computational effort
and to take into account the different importance of pixels, we propose a
lightweight convolutional neural network with a convolutional block attention
module (CBAM) for finger vein recognition, which can achieve a more accurate
capture of visual structures through an attention mechanism. First, image
sequences are fed into a lightweight convolutional neural network we designed
to improve visual features. Afterwards, it learns to assign feature weights in
an adaptive manner with the help of a convolutional block attention module. The
experiments are carried out on two publicly available databases and the results
demonstrate that the proposed method achieves a stable, highly accurate, and
robust performance in multimodal finger recognition.
Related papers
- Unleashing the Power of Depth and Pose Estimation Neural Networks by
Designing Compatible Endoscopic Images [12.412060445862842]
We conduct a detail analysis of the properties of endoscopic images and improve the compatibility of images and neural networks.
First, we introcude the Mask Image Modelling (MIM) module, which inputs partial image information instead of complete image information.
Second, we propose a lightweight neural network to enhance the endoscopic images, to explicitly improve the compatibility between images and neural networks.
arXiv Detail & Related papers (2023-09-14T02:19:38Z) - Rethinking Unsupervised Neural Superpixel Segmentation [6.123324869194195]
unsupervised learning for superpixel segmentation via CNNs has been studied.
We propose three key elements to improve the efficacy of such networks.
By experimenting with the BSDS500 dataset, we find evidence to the significance of our proposal.
arXiv Detail & Related papers (2022-06-21T09:30:26Z) - Understanding the Influence of Receptive Field and Network Complexity in
Neural-Network-Guided TEM Image Analysis [0.0]
We systematically examine how neural network architecture choices affect how neural networks segment in transmission electron microscopy (TEM) images.
We find that for low-resolution TEM images which rely on amplitude contrast to distinguish nanoparticles from background, the receptive field does not significantly influence segmentation performance.
On the other hand, for high-resolution TEM images which rely on a combination of amplitude and phase contrast changes to identify nanoparticles, receptive field is a key parameter for increased performance.
arXiv Detail & Related papers (2022-04-08T18:45:15Z) - Vision Transformer with Convolutions Architecture Search [72.70461709267497]
We propose an architecture search method-Vision Transformer with Convolutions Architecture Search (VTCAS)
The high-performance backbone network searched by VTCAS introduces the desirable features of convolutional neural networks into the Transformer architecture.
It enhances the robustness of the neural network for object recognition, especially in the low illumination indoor scene.
arXiv Detail & Related papers (2022-03-20T02:59:51Z) - Visual Attention Network [90.0753726786985]
We propose a novel large kernel attention (LKA) module to enable self-adaptive and long-range correlations in self-attention.
We also introduce a novel neural network based on LKA, namely Visual Attention Network (VAN)
VAN outperforms the state-of-the-art vision transformers and convolutional neural networks with a large margin in extensive experiments.
arXiv Detail & Related papers (2022-02-20T06:35:18Z) - Data-driven emergence of convolutional structure in neural networks [83.4920717252233]
We show how fully-connected neural networks solving a discrimination task can learn a convolutional structure directly from their inputs.
By carefully designing data models, we show that the emergence of this pattern is triggered by the non-Gaussian, higher-order local structure of the inputs.
arXiv Detail & Related papers (2022-02-01T17:11:13Z) - Cross-modal Attention for MRI and Ultrasound Volume Registration [10.725645523967904]
We develop a self-attention mechanism specifically for cross-modal image registration.
Our proposed cross-modal attention block effectively maps each of the features in one volume to all features in the corresponding volume.
Our experimental results demonstrate that a CNN network designed with the cross-modal attention block embedded outperforms an advanced CNN network 10 times of its size.
arXiv Detail & Related papers (2021-07-09T17:04:02Z) - Deep Features for training Support Vector Machine [16.795405355504077]
This paper develops a generic computer vision system based on features extracted from trained CNNs.
Multiple learned features are combined into a single structure to work on different image classification tasks.
arXiv Detail & Related papers (2021-04-08T03:13:09Z) - The Mind's Eye: Visualizing Class-Agnostic Features of CNNs [92.39082696657874]
We propose an approach to visually interpret CNN features given a set of images by creating corresponding images that depict the most informative features of a specific layer.
Our method uses a dual-objective activation and distance loss, without requiring a generator network nor modifications to the original model.
arXiv Detail & Related papers (2021-01-29T07:46:39Z) - Channel-Level Variable Quantization Network for Deep Image Compression [50.3174629451739]
We propose a channel-level variable quantization network to dynamically allocate more convolutions for significant channels and withdraws for negligible channels.
Our method achieves superior performance and can produce much better visual reconstructions.
arXiv Detail & Related papers (2020-07-15T07:20:39Z) - Ventral-Dorsal Neural Networks: Object Detection via Selective Attention [51.79577908317031]
We propose a new framework called Ventral-Dorsal Networks (VDNets)
Inspired by the structure of the human visual system, we propose the integration of a "Ventral Network" and a "Dorsal Network"
Our experimental results reveal that the proposed method outperforms state-of-the-art object detection approaches.
arXiv Detail & Related papers (2020-05-15T23:57:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.