Video Face Manipulation Detection Through Ensemble of CNNs
- URL: http://arxiv.org/abs/2004.07676v1
- Date: Thu, 16 Apr 2020 14:19:40 GMT
- Title: Video Face Manipulation Detection Through Ensemble of CNNs
- Authors: Nicol\`o Bonettini, Edoardo Daniele Cannas, Sara Mandelli, Luca Bondi,
Paolo Bestagini, Stefano Tubaro
- Abstract summary: We tackle the problem of face manipulation detection in video sequences targeting modern facial manipulation techniques.
In particular, we study the ensembling of different trained Convolutional Neural Network (CNN) models.
We show that combining these networks leads to promising face manipulation detection results on two publicly available datasets.
- Score: 17.051112469244778
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the last few years, several techniques for facial manipulation in videos
have been successfully developed and made available to the masses (i.e.,
FaceSwap, deepfake, etc.). These methods enable anyone to easily edit faces in
video sequences with incredibly realistic results and a very little effort.
Despite the usefulness of these tools in many fields, if used maliciously, they
can have a significantly bad impact on society (e.g., fake news spreading,
cyber bullying through fake revenge porn). The ability of objectively detecting
whether a face has been manipulated in a video sequence is then a task of
utmost importance. In this paper, we tackle the problem of face manipulation
detection in video sequences targeting modern facial manipulation techniques.
In particular, we study the ensembling of different trained Convolutional
Neural Network (CNN) models. In the proposed solution, different models are
obtained starting from a base network (i.e., EfficientNetB4) making use of two
different concepts: (i) attention layers; (ii) siamese training. We show that
combining these networks leads to promising face manipulation detection results
on two publicly available datasets with more than 119000 videos.
Related papers
- Deepfake detection in videos with multiple faces using geometric-fakeness features [79.16635054977068]
Deepfakes of victims or public figures can be used by fraudsters for blackmailing, extorsion and financial fraud.
In our research we propose to use geometric-fakeness features (GFF) that characterize a dynamic degree of a face presence in a video.
We employ our approach to analyze videos with multiple faces that are simultaneously present in a video.
arXiv Detail & Related papers (2024-10-10T13:10:34Z) - Face Forgery Detection Based on Facial Region Displacement Trajectory
Series [10.338298543908339]
We develop a method for detecting manipulated videos based on the trajectory of the facial region displacement.
This information was used to construct a network for exposing multidimensional artifacts in the trajectory sequences of manipulated videos.
arXiv Detail & Related papers (2022-12-07T14:47:54Z) - Video Manipulations Beyond Faces: A Dataset with Human-Machine Analysis [60.13902294276283]
We present VideoSham, a dataset consisting of 826 videos (413 real and 413 manipulated).
Many of the existing deepfake datasets focus exclusively on two types of facial manipulations -- swapping with a different subject's face or altering the existing face.
Our analysis shows that state-of-the-art manipulation detection algorithms only work for a few specific attacks and do not scale well on VideoSham.
arXiv Detail & Related papers (2022-07-26T17:39:04Z) - Convolutional Neural Network Based Partial Face Detection [0.0]
This study aims to create and enhance a machine learning model that correctly recognizes faces.
After creating and running the model, Multi-Task Convolutional Neural Network (MTCNN) achieved 96.2% best model accuracy.
arXiv Detail & Related papers (2022-06-29T01:26:40Z) - Video2StyleGAN: Encoding Video in Latent Space for Manipulation [63.03250800510085]
We propose a novel network to encode face videos into the latent space of StyleGAN for semantic face video manipulation.
Our approach can significantly outperform existing single image methods, while achieving real-time (66 fps) speed.
arXiv Detail & Related papers (2022-06-27T06:48:15Z) - Copy Motion From One to Another: Fake Motion Video Generation [53.676020148034034]
A compelling application of artificial intelligence is to generate a video of a target person performing arbitrary desired motion.
Current methods typically employ GANs with a L2 loss to assess the authenticity of the generated videos.
We propose a theoretically motivated Gromov-Wasserstein loss that facilitates learning the mapping from a pose to a foreground image.
Our method is able to generate realistic target person videos, faithfully copying complex motions from a source person.
arXiv Detail & Related papers (2022-05-03T08:45:22Z) - Audio-Visual Person-of-Interest DeepFake Detection [77.04789677645682]
The aim of this work is to propose a deepfake detector that can cope with the wide variety of manipulation methods and scenarios encountered in the real world.
We leverage a contrastive learning paradigm to learn the moving-face and audio segment embeddings that are most discriminative for each identity.
Our method can detect both single-modality (audio-only, video-only) and multi-modality (audio-video) attacks, and is robust to low-quality or corrupted videos.
arXiv Detail & Related papers (2022-04-06T20:51:40Z) - Watch Those Words: Video Falsification Detection Using Word-Conditioned
Facial Motion [82.06128362686445]
We propose a multi-modal semantic forensic approach to handle both cheapfakes and visually persuasive deepfakes.
We leverage the idea of attribution to learn person-specific biometric patterns that distinguish a given speaker from others.
Unlike existing person-specific approaches, our method is also effective against attacks that focus on lip manipulation.
arXiv Detail & Related papers (2021-12-21T01:57:04Z) - Detecting Deepfake Videos Using Euler Video Magnification [1.8506048493564673]
Deepfake videos are manipulating videos using advanced machine learning techniques.
In this paper, we examine a technique for possible identification of deepfake videos.
Our approach uses features extracted from the Euler technique to train three models to classify counterfeit and unaltered videos.
arXiv Detail & Related papers (2021-01-27T17:37:23Z) - Detecting Deep-Fake Videos from Appearance and Behavior [0.0]
We describe a biometric-based forensic technique for detecting face-swap deep fakes.
We show the efficacy of this approach across several large-scale video datasets.
arXiv Detail & Related papers (2020-04-29T21:38:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.