BPFNet: A Unified Framework for Bimodal Palmprint Alignment and Fusion
- URL: http://arxiv.org/abs/2110.01179v1
- Date: Mon, 4 Oct 2021 04:30:36 GMT
- Title: BPFNet: A Unified Framework for Bimodal Palmprint Alignment and Fusion
- Authors: Zhaoqun Li, Xu Liang, Dandan Fan, Jinxing Li, David Zhang
- Abstract summary: Bimodal palmprint recognition leverages palmprint and palm vein images simultaneously.
In this paper, we propose Bimodal Palmprint Fusion Network (BPFNet) which focuses on ROI localization, alignment and bimodal image fusion.
- Score: 26.770714818092774
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Bimodal palmprint recognition leverages palmprint and palm vein images
simultaneously,which achieves high accuracy by multi-model information fusion
and has strong anti-falsification property. In the recognition pipeline, the
detection of palm and the alignment of region-of-interest (ROI) are two crucial
steps for accurate matching. Most existing methods localize palm ROI by
keypoint detection algorithms, however the intrinsic difficulties of keypoint
detection tasks make the results unsatisfactory. Besides, the ROI alignment and
fusion algorithms at image-level are not fully investigaged.To bridge the gap,
in this paper, we propose Bimodal Palmprint Fusion Network (BPFNet) which
focuses on ROI localization, alignment and bimodal image fusion.BPFNet is an
end-to-end framework containing two subnets: The detection network directly
regresses the palmprint ROIs based on bounding box prediction and conducts
alignment by translation estimation.In the downstream,the bimodal fusion
network implements bimodal ROI image fusion leveraging a novel proposed
cross-modal selection scheme. To show the effectiveness of BPFNet,we carry out
experiments on the large-scale touchless palmprint datasets CUHKSZ-v1 and
TongJi and the proposed method achieves state-of-the-art performances.
Related papers
- Embracing Events and Frames with Hierarchical Feature Refinement Network for Object Detection [17.406051477690134]
Event cameras output sparse and asynchronous events, providing a potential solution to solve these problems.
We propose a novel hierarchical feature refinement network for event-frame fusion.
Our method exhibits significantly better robustness when introducing 15 different corruption types to the frame images.
arXiv Detail & Related papers (2024-07-17T14:09:46Z) - From Text to Pixels: A Context-Aware Semantic Synergy Solution for
Infrared and Visible Image Fusion [66.33467192279514]
We introduce a text-guided multi-modality image fusion method that leverages the high-level semantics from textual descriptions to integrate semantics from infrared and visible images.
Our method not only produces visually superior fusion results but also achieves a higher detection mAP over existing methods, achieving state-of-the-art results.
arXiv Detail & Related papers (2023-12-31T08:13:47Z) - DPTNet: A Dual-Path Transformer Architecture for Scene Text Detection [34.42038300372715]
We propose DPTNet, a simple yet effective architecture to model the global and local information for the scene text detection task.
We propose a parallel design that integrates the convolutional network with a powerful self-attention mechanism to provide complementary clues between the attention path and convolutional path.
Our DPTNet achieves state-of-the-art results on the MSRA-TD500 dataset, and provides competitive results on other standard benchmarks in terms of both detection accuracy and speed.
arXiv Detail & Related papers (2022-08-21T12:58:45Z) - PCAM: Product of Cross-Attention Matrices for Rigid Registration of
Point Clouds [79.99653758293277]
PCAM is a neural network whose key element is a pointwise product of cross-attention matrices.
We show that PCAM achieves state-of-the-art results among methods which, like us, solve steps (a) and (b) jointly via deepnets.
arXiv Detail & Related papers (2021-10-04T09:23:27Z) - MBDF-Net: Multi-Branch Deep Fusion Network for 3D Object Detection [17.295359521427073]
We propose a Multi-Branch Deep Fusion Network (MBDF-Net) for 3D object detection.
In the first stage, our multi-branch feature extraction network utilizes Adaptive Attention Fusion modules to produce cross-modal fusion features from single-modal semantic features.
In the second stage, we use a region of interest (RoI) -pooled fusion module to generate enhanced local features for refinement.
arXiv Detail & Related papers (2021-08-29T15:40:15Z) - EPMF: Efficient Perception-aware Multi-sensor Fusion for 3D Semantic Segmentation [62.210091681352914]
We study multi-sensor fusion for 3D semantic segmentation for many applications, such as autonomous driving and robotics.
In this work, we investigate a collaborative fusion scheme called perception-aware multi-sensor fusion (PMF)
We propose a two-stream network to extract features from the two modalities separately. The extracted features are fused by effective residual-based fusion modules.
arXiv Detail & Related papers (2021-06-21T10:47:26Z) - Hierarchical Convolutional Neural Network with Feature Preservation and
Autotuned Thresholding for Crack Detection [5.735035463793008]
Drone imagery is increasingly used in automated inspection for infrastructure surface defects.
This paper proposes a deep learning approach using hierarchical convolutional neural networks with feature preservation.
The proposed technique is then applied to identify surface cracks on the surface of roads, bridges or pavements.
arXiv Detail & Related papers (2021-04-21T13:07:58Z) - Boundary-Aware Segmentation Network for Mobile and Web Applications [60.815545591314915]
Boundary-Aware Network (BASNet) is integrated with a predict-refine architecture and a hybrid loss for highly accurate image segmentation.
BASNet runs at over 70 fps on a single GPU which benefits many potential real applications.
Based on BASNet, we further developed two (close to) commercial applications: AR COPY & PASTE, in which BASNet is augmented reality for "COPY" and "PASTING" real-world objects, and OBJECT CUT, which is a web-based tool for automatic object background removal.
arXiv Detail & Related papers (2021-01-12T19:20:26Z) - Two-Stage Copy-Move Forgery Detection with Self Deep Matching and
Proposal SuperGlue [9.676233763589618]
Copy-move forgery detection identifies a tampered image by detecting pasted and source regions in the same image.
We propose a novel two-stage framework specially for copy-move forgery detection.
Proposal SuperGlue is proposed to remove false-alarmed regions and remedy incomplete regions.
arXiv Detail & Related papers (2020-12-16T02:05:55Z) - Deep Semantic Matching with Foreground Detection and Cycle-Consistency [103.22976097225457]
We address weakly supervised semantic matching based on a deep network.
We explicitly estimate the foreground regions to suppress the effect of background clutter.
We develop cycle-consistent losses to enforce the predicted transformations across multiple images to be geometrically plausible and consistent.
arXiv Detail & Related papers (2020-03-31T22:38:09Z) - DPANet: Depth Potentiality-Aware Gated Attention Network for RGB-D
Salient Object Detection [107.96418568008644]
We propose a novel network named DPANet to explicitly model the potentiality of the depth map and effectively integrate the cross-modal complementarity.
By introducing the depth potentiality perception, the network can perceive the potentiality of depth information in a learning-based manner.
arXiv Detail & Related papers (2020-03-19T07:27:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.