Mutual Information Maximization on Disentangled Representations for
Differential Morph Detection
- URL: http://arxiv.org/abs/2012.01542v1
- Date: Wed, 2 Dec 2020 21:31:02 GMT
- Title: Mutual Information Maximization on Disentangled Representations for
Differential Morph Detection
- Authors: Sobhan Soleymani, Ali Dabouei, Fariborz Taherkhani, Jeremy Dawson,
Nasser M. Nasrabadi
- Abstract summary: We present a novel differential morph detection framework, utilizing landmark and appearance disentanglement.
The proposed framework can provide state-of-the-art differential morph detection performance.
- Score: 29.51265709271036
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we present a novel differential morph detection framework,
utilizing landmark and appearance disentanglement. In our framework, the face
image is represented in the embedding domain using two disentangled but
complementary representations. The network is trained by triplets of face
images, in which the intermediate image inherits the landmarks from one image
and the appearance from the other image. This initially trained network is
further trained for each dataset using contrastive representations. We
demonstrate that, by employing appearance and landmark disentanglement, the
proposed framework can provide state-of-the-art differential morph detection
performance. This functionality is achieved by the using distances in landmark,
appearance, and ID domains. The performance of the proposed framework is
evaluated using three morph datasets generated with different methodologies.
Related papers
- Foveation in the Era of Deep Learning [6.602118206533142]
We introduce an end-to-end differentiable foveated active vision architecture that leverages a graph convolutional network to process foveated images.
Our model learns to iteratively attend to regions of the image relevant for classification.
We find that our model outperforms a state-of-the-art CNN and foveated vision architectures of comparable parameters and a given pixel or budget.
arXiv Detail & Related papers (2023-12-03T16:48:09Z) - Cross-Image Attention for Zero-Shot Appearance Transfer [68.43651329067393]
We introduce a cross-image attention mechanism that implicitly establishes semantic correspondences across images.
We harness three mechanisms that either manipulate the noisy latent codes or the model's internal representations throughout the denoising process.
Experiments show that our method is effective across a wide range of object categories and is robust to variations in shape, size, and viewpoint.
arXiv Detail & Related papers (2023-11-06T18:33:24Z) - Face Feature Visualisation of Single Morphing Attack Detection [13.680968065638108]
This paper proposes an explainable visualisation of different face feature extraction algorithms.
It enables the detection of bona fide and morphing images for single morphing attack detection.
The visualisation may help to develop a Graphical User Interface for border policies.
arXiv Detail & Related papers (2023-04-25T17:51:23Z) - Cross-Modal Fusion Distillation for Fine-Grained Sketch-Based Image
Retrieval [55.21569389894215]
We propose a cross-attention framework for Vision Transformers (XModalViT) that fuses modality-specific information instead of discarding them.
Our framework first maps paired datapoints from the individual photo and sketch modalities to fused representations that unify information from both modalities.
We then decouple the input space of the aforementioned modality fusion network into independent encoders of the individual modalities via contrastive and relational cross-modal knowledge distillation.
arXiv Detail & Related papers (2022-10-19T11:50:14Z) - Information Maximization for Extreme Pose Face Recognition [18.545778273606427]
We exploit a connection using a coupled-encoder network to project frontal/profile face images into a common embedding space.
The proposed model forces the similarity of representations in the embedding space by maximizing the mutual information between two views of the face.
arXiv Detail & Related papers (2022-09-07T20:30:06Z) - Self-Supervised Image Representation Learning with Geometric Set
Consistency [50.12720780102395]
We propose a method for self-supervised image representation learning under the guidance of 3D geometric consistency.
Specifically, we introduce 3D geometric consistency into a contrastive learning framework to enforce the feature consistency within image views.
arXiv Detail & Related papers (2022-03-29T08:57:33Z) - Ablation study of self-supervised learning for image classification [0.0]
This project focuses on the self-supervised training of convolutional neural networks (CNNs) and transformer networks for the task of image recognition.
A simple siamese network with different backbones is used in order to maximize the similarity of two augmented transformed images from the same source image.
arXiv Detail & Related papers (2021-12-04T09:59:01Z) - Shelf-Supervised Mesh Prediction in the Wild [54.01373263260449]
We propose a learning-based approach to infer 3D shape and pose of object from a single image.
We first infer a volumetric representation in a canonical frame, along with the camera pose.
The coarse volumetric prediction is then converted to a mesh-based representation, which is further refined in the predicted camera frame.
arXiv Detail & Related papers (2021-02-11T18:57:10Z) - Unsupervised Learning of Landmarks based on Inter-Intra Subject
Consistencies [72.67344725725961]
We present a novel unsupervised learning approach to image landmark discovery by incorporating the inter-subject landmark consistencies on facial images.
This is achieved via an inter-subject mapping module that transforms original subject landmarks based on an auxiliary subject-related structure.
To recover from the transformed images back to the original subject, the landmark detector is forced to learn spatial locations that contain the consistent semantic meanings both for the paired intra-subject images and between the paired inter-subject images.
arXiv Detail & Related papers (2020-04-16T20:38:16Z) - Cross-domain Correspondence Learning for Exemplar-based Image
Translation [59.35767271091425]
We present a framework for exemplar-based image translation, which synthesizes a photo-realistic image from the input in a distinct domain.
The output has the style (e.g., color, texture) in consistency with the semantically corresponding objects in the exemplar.
We show that our method is superior to state-of-the-art methods in terms of image quality significantly.
arXiv Detail & Related papers (2020-04-12T09:10:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.