Teacher-Student Adversarial Depth Hallucination to Improve Face
Recognition
- URL: http://arxiv.org/abs/2104.02424v1
- Date: Tue, 6 Apr 2021 11:07:02 GMT
- Title: Teacher-Student Adversarial Depth Hallucination to Improve Face
Recognition
- Authors: Hardik Uppal, Alireza Sepas-Moghaddam, Michael Greenspan, Ali Etemad
- Abstract summary: We present the Teacher-Student Generative Adversarial Network (TS-GAN) to generate depth images from a single RGB image.
For our method to generalize well across unseen datasets, we design two components in the architecture, a teacher and a student.
The fully trained shared generator can then be used in runtime to hallucinate depth from RGB for downstream applications such as face recognition.
- Score: 11.885178256393893
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present the Teacher-Student Generative Adversarial Network (TS-GAN) to
generate depth images from a single RGB image in order to boost the recognition
accuracy of face recognition (FR) systems. For our method to generalize well
across unseen datasets, we design two components in the architecture, a teacher
and a student. The teacher, which itself consists of a generator and a
discriminator, learns a latent mapping between input RGB and paired depth
images in a supervised fashion. The student, which consists of two generators
(one shared with the teacher) and a discriminator, learns from new RGB data
with no available paired depth information, for improved generalization. The
fully trained shared generator can then be used in runtime to hallucinate depth
from RGB for downstream applications such as face recognition. We perform
rigorous experiments to show the superiority of TS-GAN over other methods in
generating synthetic depth images. Moreover, face recognition experiments
demonstrate that our hallucinated depth along with the input RGB images boosts
performance across various architectures when compared to a single RGB modality
by average values of +1.2%, +2.6%, and +2.6% for IIIT-D, EURECOM, and LFW
datasets respectively.
Related papers
- Confidence-Aware RGB-D Face Recognition via Virtual Depth Synthesis [48.59382455101753]
2D face recognition encounters challenges in unconstrained environments due to varying illumination, occlusion, and pose.
Recent studies focus on RGB-D face recognition to improve robustness by incorporating depth information.
In this work, we first construct a diverse depth dataset generated by 3D Morphable Models for depth model pre-training.
Then, we propose a domain-independent pre-training framework that utilizes readily available pre-trained RGB and depth models to separately perform face recognition without needing additional paired data for retraining.
arXiv Detail & Related papers (2024-03-11T09:12:24Z) - Improving 2D face recognition via fine-level facial depth generation and
RGB-D complementary feature learning [0.8223798883838329]
We propose a fine-grained facial depth generation network and an improved multimodal complementary feature learning network.
Experiments on the Lock3DFace dataset and the IIIT-D dataset show that the proposed FFDGNet and I MCFLNet can improve the accuracy of RGB-D face recognition.
arXiv Detail & Related papers (2023-05-08T02:33:59Z) - Boosting RGB-D Saliency Detection by Leveraging Unlabeled RGB Images [89.81919625224103]
Training deep models for RGB-D salient object detection (SOD) often requires a large number of labeled RGB-D images.
We present a Dual-Semi RGB-D Salient Object Detection Network (DS-Net) to leverage unlabeled RGB images for boosting RGB-D saliency detection.
arXiv Detail & Related papers (2022-01-01T03:02:27Z) - Semantic-embedded Unsupervised Spectral Reconstruction from Single RGB
Images in the Wild [48.44194221801609]
We propose a new lightweight and end-to-end learning-based framework to tackle this challenge.
We progressively spread the differences between input RGB images and re-projected RGB images from recovered HS images via effective camera spectral response function estimation.
Our method significantly outperforms state-of-the-art unsupervised methods and even exceeds the latest supervised method under some settings.
arXiv Detail & Related papers (2021-08-15T05:19:44Z) - Cross-modality Discrepant Interaction Network for RGB-D Salient Object
Detection [78.47767202232298]
We propose a novel Cross-modality Discrepant Interaction Network (CDINet) for RGB-D SOD.
Two components are designed to implement the effective cross-modality interaction.
Our network outperforms $15$ state-of-the-art methods both quantitatively and qualitatively.
arXiv Detail & Related papers (2021-08-04T11:24:42Z) - NormalGAN: Learning Detailed 3D Human from a Single RGB-D Image [34.79657678041356]
We propose a fast adversarial learning-based method to reconstruct the complete and detailed 3D human from a single RGB-D image.
Given a consumer RGB-D sensor, NormalGAN can generate the complete and detailed 3D human reconstruction results in 20 fps.
arXiv Detail & Related papers (2020-07-30T09:35:46Z) - Bi-directional Cross-Modality Feature Propagation with
Separation-and-Aggregation Gate for RGB-D Semantic Segmentation [59.94819184452694]
Depth information has proven to be a useful cue in the semantic segmentation of RGBD images for providing a geometric counterpart to the RGB representation.
Most existing works simply assume that depth measurements are accurate and well-aligned with the RGB pixels and models the problem as a cross-modal feature fusion.
In this paper, we propose a unified and efficient Crossmodality Guided to not only effectively recalibrate RGB feature responses, but also to distill accurate depth information via multiple stages and aggregate the two recalibrated representations alternatively.
arXiv Detail & Related papers (2020-07-17T18:35:24Z) - RGB-IR Cross-modality Person ReID based on Teacher-Student GAN Model [20.70796497371778]
We propose a Teacher-Student GAN model (TS-GAN) to adopt different domains and guide the ReID backbone to learn better ReID information.
Unlike other GAN based models, the proposed model only needs the backbone module at the test stage, making it more efficient and resource-saving.
arXiv Detail & Related papers (2020-07-15T02:58:46Z) - Two-Level Attention-based Fusion Learning for RGB-D Face Recognition [21.735238213921804]
A novel attention aware method is proposed to fuse two image modalities, RGB and depth, for enhanced RGB-D facial recognition.
The proposed method first extracts features from both modalities using a convolutional feature extractor.
These features are then fused using a two-layer attention mechanism.
arXiv Detail & Related papers (2020-02-29T03:18:52Z) - Supervised and Unsupervised Learning of Parameterized Color Enhancement [112.88623543850224]
We tackle the problem of color enhancement as an image translation task using both supervised and unsupervised learning.
We achieve state-of-the-art results compared to both supervised (paired data) and unsupervised (unpaired data) image enhancement methods on the MIT-Adobe FiveK benchmark.
We show the generalization capability of our method, by applying it on photos from the early 20th century and to dark video frames.
arXiv Detail & Related papers (2019-12-30T13:57:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.