LibreFace: An Open-Source Toolkit for Deep Facial Expression Analysis
- URL: http://arxiv.org/abs/2308.10713v2
- Date: Thu, 24 Aug 2023 03:46:53 GMT
- Title: LibreFace: An Open-Source Toolkit for Deep Facial Expression Analysis
- Authors: Di Chang, Yufeng Yin, Zongjian Li, Minh Tran, Mohammad Soleymani
- Abstract summary: We introduce LibreFace, an open-source toolkit for facial expression analysis.
It offers real-time and offline analysis of facial behavior through deep learning models.
Our model also demonstrates competitive performance to state-of-the-art facial expression analysis methods.
- Score: 7.185007035384591
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Facial expression analysis is an important tool for human-computer
interaction. In this paper, we introduce LibreFace, an open-source toolkit for
facial expression analysis. This open-source toolbox offers real-time and
offline analysis of facial behavior through deep learning models, including
facial action unit (AU) detection, AU intensity estimation, and facial
expression recognition. To accomplish this, we employ several techniques,
including the utilization of a large-scale pre-trained network, feature-wise
knowledge distillation, and task-specific fine-tuning. These approaches are
designed to effectively and accurately analyze facial expressions by leveraging
visual information, thereby facilitating the implementation of real-time
interactive applications. In terms of Action Unit (AU) intensity estimation, we
achieve a Pearson Correlation Coefficient (PCC) of 0.63 on DISFA, which is 7%
higher than the performance of OpenFace 2.0 while maintaining highly-efficient
inference that runs two times faster than OpenFace 2.0. Despite being compact,
our model also demonstrates competitive performance to state-of-the-art facial
expression analysis methods on AffecNet, FFHQ, and RAF-DB. Our code will be
released at https://github.com/ihp-lab/LibreFace
Related papers
- UniLearn: Enhancing Dynamic Facial Expression Recognition through Unified Pre-Training and Fine-Tuning on Images and Videos [83.48170683672427]
UniLearn is a unified learning paradigm that integrates static facial expression recognition data to enhance DFER task.
UniLearn consistently state-of-the-art performance on FERV39K, MAFW, and DFEW benchmarks, with weighted average recall (WAR) of 53.65%, 58.44%, and 76.68%, respectively.
arXiv Detail & Related papers (2024-09-10T01:57:57Z) - Task-adaptive Q-Face [75.15668556061772]
We propose a novel task-adaptive multi-task face analysis method named as Q-Face.
Q-Face simultaneously performs multiple face analysis tasks with a unified model.
Our method achieves state-of-the-art performance on face expression recognition, action unit detection, face attribute analysis, age estimation, and face pose estimation.
arXiv Detail & Related papers (2024-05-15T03:13:11Z) - A Generative Framework for Self-Supervised Facial Representation Learning [18.094262972295702]
Self-supervised representation learning has gained increasing attention for strong generalization ability without relying on paired datasets.
Self-supervised facial representation learning remains unsolved due to the coupling of facial identities, expressions, and external factors like pose and light.
We propose LatentFace, a novel generative framework for self-supervised facial representations.
arXiv Detail & Related papers (2023-09-15T09:34:05Z) - SwinFace: A Multi-task Transformer for Face Recognition, Expression
Recognition, Age Estimation and Attribute Estimation [60.94239810407917]
This paper presents a multi-purpose algorithm for simultaneous face recognition, facial expression recognition, age estimation, and face attribute estimation based on a single Swin Transformer.
To address the conflicts among multiple tasks, a Multi-Level Channel Attention (MLCA) module is integrated into each task-specific analysis.
Experiments show that the proposed model has a better understanding of the face and achieves excellent performance for all tasks.
arXiv Detail & Related papers (2023-08-22T15:38:39Z) - Frame-level Prediction of Facial Expressions, Valence, Arousal and
Action Units for Mobile Devices [7.056222499095849]
We propose the novel frame-level emotion recognition algorithm by extracting facial features with the single EfficientNet model pre-trained on AffectNet.
Our approach may be implemented even for video analytics on mobile devices.
arXiv Detail & Related papers (2022-03-25T03:53:27Z) - AFFDEX 2.0: A Real-Time Facial Expression Analysis Toolkit [1.076535942003539]
AFFDEX 2.0 is a toolkit for analyzing facial expressions in the wild.
It can estimate the 3D head pose, detect facial Action Units (AUs), recognize basic emotions and 2 new emotional states (sentimentality and confusion)
AFFDEX 2.0 can process multiple faces in real time, and is working across the Windows and Linux platforms.
arXiv Detail & Related papers (2022-02-24T12:27:49Z) - FP-Age: Leveraging Face Parsing Attention for Facial Age Estimation in
the Wild [50.8865921538953]
We propose a method to explicitly incorporate facial semantics into age estimation.
We design a face parsing-based network to learn semantic information at different scales.
We show that our method consistently outperforms all existing age estimation methods.
arXiv Detail & Related papers (2021-06-21T14:31:32Z) - A Multi-resolution Approach to Expression Recognition in the Wild [9.118706387430883]
We propose a multi-resolution approach to solve the Facial Expression Recognition task.
We ground our intuition on the observation that often faces images are acquired at different resolutions.
To our aim, we use a ResNet-like architecture, equipped with Squeeze-and-Excitation blocks, trained on the Affect-in-the-Wild 2 dataset.
arXiv Detail & Related papers (2021-03-09T21:21:02Z) - Unsupervised Learning Facial Parameter Regressor for Action Unit
Intensity Estimation via Differentiable Renderer [51.926868759681014]
We present a framework to predict the facial parameters based on a bone-driven face model (BDFM) under different views.
The proposed framework consists of a feature extractor, a generator, and a facial parameter regressor.
arXiv Detail & Related papers (2020-08-20T09:49:13Z) - Learning to Augment Expressions for Few-shot Fine-grained Facial
Expression Recognition [98.83578105374535]
We present a novel Fine-grained Facial Expression Database - F2ED.
It includes more than 200k images with 54 facial expressions from 119 persons.
Considering the phenomenon of uneven data distribution and lack of samples is common in real-world scenarios, we evaluate several tasks of few-shot expression learning.
We propose a unified task-driven framework - Compositional Generative Adversarial Network (Comp-GAN) learning to synthesize facial images.
arXiv Detail & Related papers (2020-01-17T03:26:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.