Information Maximization for Extreme Pose Face Recognition
- URL: http://arxiv.org/abs/2209.03456v1
- Date: Wed, 7 Sep 2022 20:30:06 GMT
- Title: Information Maximization for Extreme Pose Face Recognition
- Authors: Mohammad Saeed Ebrahimi Saadabadi, Sahar Rahimi Malakshan, Sobhan
Soleymani, Moktari Mostofa, and Nasser M. Nasrabadi
- Abstract summary: We exploit a connection using a coupled-encoder network to project frontal/profile face images into a common embedding space.
The proposed model forces the similarity of representations in the embedding space by maximizing the mutual information between two views of the face.
- Score: 18.545778273606427
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we seek to draw connections between the frontal and profile
face images in an abstract embedding space. We exploit this connection using a
coupled-encoder network to project frontal/profile face images into a common
latent embedding space. The proposed model forces the similarity of
representations in the embedding space by maximizing the mutual information
between two views of the face. The proposed coupled-encoder benefits from three
contributions for matching faces with extreme pose disparities. First, we
leverage our pose-aware contrastive learning to maximize the mutual information
between frontal and profile representations of identities. Second, a memory
buffer, which consists of latent representations accumulated over past
iterations, is integrated into the model so it can refer to relatively much
more instances than the mini-batch size. Third, a novel pose-aware adversarial
domain adaptation method forces the model to learn an asymmetric mapping from
profile to frontal representation. In our framework, the coupled-encoder learns
to enlarge the margin between the distribution of genuine and imposter faces,
which results in high mutual information between different views of the same
identity. The effectiveness of the proposed model is investigated through
extensive experiments, evaluations, and ablation studies on four benchmark
datasets, and comparison with the compelling state-of-the-art algorithms.
Related papers
- DiHuR: Diffusion-Guided Generalizable Human Reconstruction [51.31232435994026]
We introduce DiHuR, a Diffusion-guided model for generalizable Human 3D Reconstruction and view synthesis from sparse, minimally overlapping images.
Our method integrates two key priors in a coherent manner: the prior from generalizable feed-forward models and the 2D diffusion prior, and it requires only multi-view image training, without 3D supervision.
arXiv Detail & Related papers (2024-11-16T03:52:23Z) - DreamMover: Leveraging the Prior of Diffusion Models for Image Interpolation with Large Motion [35.60459492849359]
We study the problem of generating intermediate images from image pairs with large motion.
Due to the large motion, the intermediate semantic information may be absent in input images.
We propose DreamMover, a novel image framework with three main components.
arXiv Detail & Related papers (2024-09-15T04:09:12Z) - Towards Self-Supervised FG-SBIR with Unified Sample Feature Alignment and Multi-Scale Token Recycling [11.129453244307369]
FG-SBIR aims to minimize the distance between sketches and corresponding images in the embedding space.
We propose an effective approach to narrow the gap between the two domains.
It mainly facilitates unified mutual information sharing both intra- and inter-samples.
arXiv Detail & Related papers (2024-06-17T13:49:12Z) - Explicit Correspondence Matching for Generalizable Neural Radiance
Fields [49.49773108695526]
We present a new NeRF method that is able to generalize to new unseen scenarios and perform novel view synthesis with as few as two source views.
The explicit correspondence matching is quantified with the cosine similarity between image features sampled at the 2D projections of a 3D point on different views.
Our method achieves state-of-the-art results on different evaluation settings, with the experiments showing a strong correlation between our learned cosine feature similarity and volume density.
arXiv Detail & Related papers (2023-04-24T17:46:01Z) - Semantic-aware One-shot Face Re-enactment with Dense Correspondence
Estimation [100.60938767993088]
One-shot face re-enactment is a challenging task due to the identity mismatch between source and driving faces.
This paper proposes to use 3D Morphable Model (3DMM) for explicit facial semantic decomposition and identity disentanglement.
arXiv Detail & Related papers (2022-11-23T03:02:34Z) - Unifying Flow, Stereo and Depth Estimation [121.54066319299261]
We present a unified formulation and model for three motion and 3D perception tasks.
We formulate all three tasks as a unified dense correspondence matching problem.
Our model naturally enables cross-task transfer since the model architecture and parameters are shared across tasks.
arXiv Detail & Related papers (2022-11-10T18:59:54Z) - Mutual Information Maximization on Disentangled Representations for
Differential Morph Detection [29.51265709271036]
We present a novel differential morph detection framework, utilizing landmark and appearance disentanglement.
The proposed framework can provide state-of-the-art differential morph detection performance.
arXiv Detail & Related papers (2020-12-02T21:31:02Z) - Face Frontalization Based on Robustly Fitting a Deformable Shape Model
to 3D Landmarks [24.07648367866321]
Face frontalization consists of a frontally-viewed face from an arbitrarily-viewed one.
The main contribution of this paper is a robust face alignment method that enables pixel-to-pixel warping.
An important merit of the proposed method is its ability to deal both with noise (small perturbations) and with outliers (large errors)
arXiv Detail & Related papers (2020-10-26T15:52:50Z) - Multi-Margin based Decorrelation Learning for Heterogeneous Face
Recognition [90.26023388850771]
This paper presents a deep neural network approach to extract decorrelation representations in a hyperspherical space for cross-domain face images.
The proposed framework can be divided into two components: heterogeneous representation network and decorrelation representation learning.
Experimental results on two challenging heterogeneous face databases show that our approach achieves superior performance on both verification and recognition tasks.
arXiv Detail & Related papers (2020-05-25T07:01:12Z) - Dual-Attention GAN for Large-Pose Face Frontalization [59.689836951934694]
We present a novel Dual-Attention Generative Adversarial Network (DA-GAN) for photo-realistic face frontalization.
Specifically, a self-attention-based generator is introduced to integrate local features with their long-range dependencies.
A novel face-attention-based discriminator is applied to emphasize local features of face regions.
arXiv Detail & Related papers (2020-02-17T20:00:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.