Unsupervised Deep Learning Image Verification Method
- URL: http://arxiv.org/abs/2312.14395v2
- Date: Tue, 6 Feb 2024 21:51:04 GMT
- Title: Unsupervised Deep Learning Image Verification Method
- Authors: Enoch Solomon, Abraham Woubie and Eyael Solomon Emiru
- Abstract summary: The proposed method achieves a relative improvement of 56% in terms of EER over the baseline system on Labeled Faces in the Wild dataset.
This has successfully narrowed down the performance gap between cosine and PLDA scoring systems.
- Score: 0.0
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Although deep learning are commonly employed for image recognition, usually
huge amount of labeled training data is required, which may not always be
readily available. This leads to a noticeable performance disparity when
compared to state-of-the-art unsupervised face verification techniques. In this
work, we propose a method to narrow this gap by leveraging an autoencoder to
convert the face image vector into a novel representation. Notably, the
autoencoder is trained to reconstruct neighboring face image vectors rather
than the original input image vectors. These neighbor face image vectors are
chosen through an unsupervised process based on the highest cosine scores with
the training face image vectors. The proposed method achieves a relative
improvement of 56\% in terms of EER over the baseline system on Labeled Faces
in the Wild (LFW) dataset. This has successfully narrowed down the performance
gap between cosine and PLDA scoring systems.
Related papers
- Co-Segmentation without any Pixel-level Supervision with Application to Large-Scale Sketch Classification [3.3104978705632777]
We propose a novel method for object co-segmentation, i.e. pixel-level localization of a common object in a set of images.
The method achieves state-of-the-art performance among methods trained with the same level of supervision.
The benefits of the proposed co-segmentation method are further demonstrated in the task of large-scale sketch recognition.
arXiv Detail & Related papers (2024-10-17T14:16:45Z) - CricaVPR: Cross-image Correlation-aware Representation Learning for Visual Place Recognition [73.51329037954866]
We propose a robust global representation method with cross-image correlation awareness for visual place recognition.
Our method uses the attention mechanism to correlate multiple images within a batch.
Our method outperforms state-of-the-art methods by a large margin with significantly less training time.
arXiv Detail & Related papers (2024-02-29T15:05:11Z) - Regressing Transformers for Data-efficient Visual Place Recognition [10.156432076272475]
This work introduces a fresh perspective by framing place recognition as a regression problem.
By optimizing image descriptors to align directly with graded similarity labels, this approach enhances ranking capabilities without expensive re-ranking.
arXiv Detail & Related papers (2024-01-29T17:04:32Z) - Improving Masked Autoencoders by Learning Where to Mask [65.89510231743692]
Masked image modeling is a promising self-supervised learning method for visual data.
We present AutoMAE, a framework that uses Gumbel-Softmax to interlink an adversarially-trained mask generator and a mask-guided image modeling process.
In our experiments, AutoMAE is shown to provide effective pretraining models on standard self-supervised benchmarks and downstream tasks.
arXiv Detail & Related papers (2023-03-12T05:28:55Z) - Masked Autoencoders Are Scalable Vision Learners [60.97703494764904]
Masked autoencoders (MAE) are scalable self-supervised learners for computer vision.
Our MAE approach is simple: we mask random patches of the input image and reconstruct the missing pixels.
Coupling these two designs enables us to train large models efficiently and effectively.
arXiv Detail & Related papers (2021-11-11T18:46:40Z) - Inverse Problems Leveraging Pre-trained Contrastive Representations [88.70821497369785]
We study a new family of inverse problems for recovering representations of corrupted data.
We propose a supervised inversion method that uses a contrastive objective to obtain excellent representations for highly corrupted images.
Our method outperforms end-to-end baselines even with a fraction of the labeled data in a wide range of forward operators.
arXiv Detail & Related papers (2021-10-14T15:06:30Z) - Self-supervised Product Quantization for Deep Unsupervised Image
Retrieval [21.99902461562925]
Supervised deep learning-based hash and vector quantization are enabling fast and large-scale image retrieval systems.
We propose the first deep unsupervised image retrieval method dubbed Self-supervised Product Quantization (SPQ) network, which is label-free and trained in a self-supervised manner.
Our method analyzes the image contents to extract descriptive features, allowing us to understand image representations for accurate retrieval.
arXiv Detail & Related papers (2021-09-06T05:02:34Z) - AugNet: End-to-End Unsupervised Visual Representation Learning with
Image Augmentation [3.6790362352712873]
We propose AugNet, a new deep learning training paradigm to learn image features from a collection of unlabeled pictures.
Our experiments demonstrate that the method is able to represent the image in low dimensional space.
Unlike many deep-learning-based image retrieval algorithms, our approach does not require access to external annotated datasets.
arXiv Detail & Related papers (2021-06-11T09:02:30Z) - CutPaste: Self-Supervised Learning for Anomaly Detection and
Localization [59.719925639875036]
We propose a framework for building anomaly detectors using normal training data only.
We first learn self-supervised deep representations and then build a generative one-class classifier on learned representations.
Our empirical study on MVTec anomaly detection dataset demonstrates the proposed algorithm is general to be able to detect various types of real-world defects.
arXiv Detail & Related papers (2021-04-08T19:04:55Z) - G-SimCLR : Self-Supervised Contrastive Learning with Guided Projection
via Pseudo Labelling [0.8164433158925593]
In computer vision, it is evident that deep neural networks perform better in a supervised setting with a large amount of labeled data.
In this work, we propose that, with the normalized temperature-scaled cross-entropy (NT-Xent) loss function, it is beneficial to not have images of the same category in the same batch.
We use the latent space representation of a denoising autoencoder trained on the unlabeled dataset and cluster them with k-means to obtain pseudo labels.
arXiv Detail & Related papers (2020-09-25T02:25:37Z) - Distilling Localization for Self-Supervised Representation Learning [82.79808902674282]
Contrastive learning has revolutionized unsupervised representation learning.
Current contrastive models are ineffective at localizing the foreground object.
We propose a data-driven approach for learning in variance to backgrounds.
arXiv Detail & Related papers (2020-04-14T16:29:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.