Minutiae-Guided Fingerprint Embeddings via Vision Transformers
- URL: http://arxiv.org/abs/2210.13994v2
- Date: Wed, 26 Oct 2022 03:26:53 GMT
- Title: Minutiae-Guided Fingerprint Embeddings via Vision Transformers
- Authors: Steven A. Grosz, Joshua J. Engelsma, Rajeev Ranjan, Naveen
Ramakrishnan, Manoj Aggarwal, Gerard G. Medioni, Anil K. Jain
- Abstract summary: We learn a discriminative fixed-length fingerprint embedding using a Vision Transformer (ViT)
We show that by fusing embeddings learned by CNNs and ViTs we can reach near parity with a commercial state-of-the-art (SOTA) matcher.
Our fixed-length embeddings can be matched orders of magnitude faster than the commercial system.
- Score: 34.516748746229325
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Minutiae matching has long dominated the field of fingerprint recognition.
However, deep networks can be used to extract fixed-length embeddings from
fingerprints. To date, the few studies that have explored the use of CNN
architectures to extract such embeddings have shown extreme promise. Inspired
by these early works, we propose the first use of a Vision Transformer (ViT) to
learn a discriminative fixed-length fingerprint embedding. We further
demonstrate that by guiding the ViT to focus in on local, minutiae related
features, we can boost the recognition performance. Finally, we show that by
fusing embeddings learned by CNNs and ViTs we can reach near parity with a
commercial state-of-the-art (SOTA) matcher. In particular, we obtain a
TAR=94.23% @ FAR=0.1% on the NIST SD 302 public-domain dataset, compared to a
SOTA commercial matcher which obtains TAR=96.71% @ FAR=0.1%. Additionally, our
fixed-length embeddings can be matched orders of magnitude faster than the
commercial system (2.5 million matches/second compared to 50K matches/second).
We make our code and models publicly available to encourage further research on
this topic: https://github.com/tba.
Related papers
- Token Pruning using a Lightweight Background Aware Vision Transformer [0.6856888934092934]
Token pruning reduces the number of input tokens to the ViT based on importance criteria of each token.
Background tokens can be pruned completely or partially before feeding to a ViT based object detector.
We show a 2 layer BAViT-small model as pre-processor to YOLOS can increase the throughput by 30% - 40% with a mAP drop of 3% without any sparse fine-tuning.
arXiv Detail & Related papers (2024-10-12T01:44:54Z) - AiluRus: A Scalable ViT Framework for Dense Prediction [95.1313839257891]
Vision transformers (ViTs) have emerged as a prevalent architecture for vision tasks owing to their impressive performance.
We propose to apply adaptive resolution for different regions in the image according to their importance.
We evaluate our proposed method on three different datasets and observe promising performance.
arXiv Detail & Related papers (2023-11-02T12:48:43Z) - AFR-Net: Attention-Driven Fingerprint Recognition Network [47.87570819350573]
We improve initial studies on the use of vision transformers (ViT) for biometric recognition, including fingerprint recognition.
We propose a realignment strategy using local embeddings extracted from intermediate feature maps within the networks to refine the global embeddings in low certainty situations.
This strategy can be applied as a wrapper to any existing deep learning network (including attention-based, CNN-based, or both) to boost its performance.
arXiv Detail & Related papers (2022-11-25T05:10:39Z) - Auto-scaling Vision Transformers without Training [84.34662535276898]
We propose As-ViT, an auto-scaling framework for Vision Transformers (ViTs) without training.
As-ViT automatically discovers and scales up ViTs in an efficient and principled manner.
As a unified framework, As-ViT achieves strong performance on classification and detection.
arXiv Detail & Related papers (2022-02-24T06:30:55Z) - PrintsGAN: Synthetic Fingerprint Generator [39.804969475699345]
PrintsGAN is a synthetic fingerprint generator capable of generating unique fingerprints along with multiple impressions for a given fingerprint.
We show the utility of the PrintsGAN generated by training a deep network to extract a fixed-length embedding from a fingerprint.
arXiv Detail & Related papers (2022-01-10T22:25:10Z) - Learning Tracking Representations via Dual-Branch Fully Transformer
Networks [82.21771581817937]
We present a Siamese-like Dual-branch network based on solely Transformers for tracking.
We extract a feature vector for each patch based on its matching results with others within an attention window.
The method achieves better or comparable results as the best-performing methods.
arXiv Detail & Related papers (2021-12-05T13:44:33Z) - Self-slimmed Vision Transformer [52.67243496139175]
Vision transformers (ViTs) have become the popular structures and outperformed convolutional neural networks (CNNs) on various vision tasks.
We propose a generic self-slimmed learning approach for vanilla ViTs, namely SiT.
Specifically, we first design a novel Token Slimming Module (TSM), which can boost the inference efficiency of ViTs.
arXiv Detail & Related papers (2021-11-24T16:48:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.