Deep Learning-based Single Image Face Depth Data Enhancement
- URL: http://arxiv.org/abs/2006.11091v3
- Date: Tue, 27 Jul 2021 09:09:50 GMT
- Title: Deep Learning-based Single Image Face Depth Data Enhancement
- Authors: Torsten Schlett, Christian Rathgeb, Christoph Busch
- Abstract summary: This work proposes a deep learning face depth enhancement method in this context.
Deep learning enhancers yield noticeably better results than the tested preexisting enhancers.
- Score: 15.41435352543715
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Face recognition can benefit from the utilization of depth data captured
using low-cost cameras, in particular for presentation attack detection
purposes. Depth video output from these capture devices can however contain
defects such as holes or general depth inaccuracies. This work proposes a deep
learning face depth enhancement method in this context of facial biometrics,
which adds a security aspect to the topic. U-Net-like architectures are
utilized, and the networks are compared against hand-crafted enhancer types, as
well as a similar depth enhancer network from related work trained for an
adjacent application scenario. All tested enhancer types exclusively use depth
data as input, which differs from methods that enhance depth based on
additional input data such as visible light color images. Synthetic face depth
ground truth images and degraded forms thereof are created with help of PRNet,
to train multiple deep learning enhancer models with different network sizes
and training configurations. Evaluations are carried out on the synthetic data,
on Kinect v1 images from the KinectFaceDB, and on in-house RealSense D435
images. These evaluations include an assessment of the falsification for
occluded face depth input, which is relevant to biometric security. The
proposed deep learning enhancers yield noticeably better results than the
tested preexisting enhancers, without overly falsifying depth data when
non-face input is provided, and are shown to reduce the error of a simple
landmark-based PAD method.
Related papers
- Robust Depth Enhancement via Polarization Prompt Fusion Tuning [112.88371907047396]
We present a framework that leverages polarization imaging to improve inaccurate depth measurements from various depth sensors.
Our method first adopts a learning-based strategy where a neural network is trained to estimate a dense and complete depth map from polarization data and a sensor depth map from different sensors.
To further improve the performance, we propose a Polarization Prompt Fusion Tuning (PPFT) strategy to effectively utilize RGB-based models pre-trained on large-scale datasets.
arXiv Detail & Related papers (2024-04-05T17:55:33Z) - Depth Map Denoising Network and Lightweight Fusion Network for Enhanced
3D Face Recognition [61.27785140017464]
We introduce an innovative Depth map denoising network (DMDNet) based on the Denoising Implicit Image Function (DIIF) to reduce noise.
We further design a powerful recognition network called Lightweight Depth and Normal Fusion network (LDNFNet) to learn unique and complementary features between different modalities.
arXiv Detail & Related papers (2024-01-01T10:46:42Z) - DeepFidelity: Perceptual Forgery Fidelity Assessment for Deepfake
Detection [67.3143177137102]
Deepfake detection refers to detecting artificially generated or edited faces in images or videos.
We propose a novel Deepfake detection framework named DeepFidelity to adaptively distinguish real and fake faces.
arXiv Detail & Related papers (2023-12-07T07:19:45Z) - Exploring Depth Information for Face Manipulation Detection [25.01910127502075]
We propose a Face Depth Map Transformer (FDMT) to estimate the face depth map patch by patch from a RGB face image.
The estimated face depth map is then considered as auxiliary information to be integrated with the backbone features.
arXiv Detail & Related papers (2022-12-29T09:00:22Z) - Deep Convolutional Pooling Transformer for Deepfake Detection [54.10864860009834]
We propose a deep convolutional Transformer to incorporate decisive image features both locally and globally.
Specifically, we apply convolutional pooling and re-attention to enrich the extracted features and enhance efficacy.
The proposed solution consistently outperforms several state-of-the-art baselines on both within- and cross-dataset experiments.
arXiv Detail & Related papers (2022-09-12T15:05:41Z) - End-to-end Learning for Joint Depth and Image Reconstruction from
Diffracted Rotation [10.896567381206715]
We propose a novel end-to-end learning approach for depth from diffracted rotation.
Our approach requires a significantly less complex model and less training data, yet it is superior to existing methods in the task of monocular depth estimation.
arXiv Detail & Related papers (2022-04-14T16:14:37Z) - M2TR: Multi-modal Multi-scale Transformers for Deepfake Detection [74.19291916812921]
forged images generated by Deepfake techniques pose a serious threat to the trustworthiness of digital information.
In this paper, we aim to capture the subtle manipulation artifacts at different scales for Deepfake detection.
We introduce a high-quality Deepfake dataset, SR-DF, which consists of 4,000 DeepFake videos generated by state-of-the-art face swapping and facial reenactment methods.
arXiv Detail & Related papers (2021-04-20T05:43:44Z) - Dual Pixel Exploration: Simultaneous Depth Estimation and Image
Restoration [77.1056200937214]
We study the formation of the DP pair which links the blur and the depth information.
We propose an end-to-end DDDNet (DP-based Depth and De Network) to jointly estimate the depth and restore the image.
arXiv Detail & Related papers (2020-12-01T06:53:57Z) - SAFENet: Self-Supervised Monocular Depth Estimation with Semantic-Aware
Feature Extraction [27.750031877854717]
We propose SAFENet that is designed to leverage semantic information to overcome the limitations of the photometric loss.
Our key idea is to exploit semantic-aware depth features that integrate the semantic and geometric knowledge.
Experiments on KITTI dataset demonstrate that our methods compete or even outperform the state-of-the-art methods.
arXiv Detail & Related papers (2020-10-06T17:22:25Z) - Do We Need Depth in State-Of-The-Art Face Authentication? [8.755493949976492]
We introduce a novel method that learns to recognize faces from stereo camera systems without the need to explicitly compute the facial surface or depth map.
The raw face stereo images along with the location in the image from which the face is extracted allow the proposed CNN to improve the recognition task.
We demonstrate that the suggested method outperforms both single-image and explicit depth based methods on large-scale benchmarks.
arXiv Detail & Related papers (2020-03-24T14:51:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.