Deep Collaborative Multi-Modal Learning for Unsupervised Kinship
Estimation
- URL: http://arxiv.org/abs/2109.02804v1
- Date: Tue, 7 Sep 2021 01:34:51 GMT
- Title: Deep Collaborative Multi-Modal Learning for Unsupervised Kinship
Estimation
- Authors: Guan-Nan Dong, Chi-Man Pun, Zheng Zhang
- Abstract summary: Kinship verification is a long-standing research challenge in computer vision.
We propose a novel deep collaborative multi-modal learning (DCML) to integrate the underlying information presented in facial properties.
Our DCML method is always superior to some state-of-the-art kinship verification methods.
- Score: 53.62256887837659
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Kinship verification is a long-standing research challenge in computer
vision. The visual differences presented to the face have a significant effect
on the recognition capabilities of the kinship systems. We argue that
aggregating multiple visual knowledge can better describe the characteristics
of the subject for precise kinship identification. Typically, the age-invariant
features can represent more natural facial details. Such age-related
transformations are essential for face recognition due to the biological
effects of aging. However, the existing methods mainly focus on employing the
single-view image features for kinship identification, while more meaningful
visual properties such as race and age are directly ignored in the feature
learning step. To this end, we propose a novel deep collaborative multi-modal
learning (DCML) to integrate the underlying information presented in facial
properties in an adaptive manner to strengthen the facial details for effective
unsupervised kinship verification. Specifically, we construct a well-designed
adaptive feature fusion mechanism, which can jointly leverage the complementary
properties from different visual perspectives to produce composite features and
draw greater attention to the most informative components of spatial feature
maps. Particularly, an adaptive weighting strategy is developed based on a
novel attention mechanism, which can enhance the dependencies between different
properties by decreasing the information redundancy in channels in a
self-adaptive manner. To validate the effectiveness of the proposed method,
extensive experimental evaluations conducted on four widely-used datasets show
that our DCML method is always superior to some state-of-the-art kinship
verification methods.
Related papers
- High-Discriminative Attribute Feature Learning for Generalized Zero-Shot Learning [54.86882315023791]
We propose an innovative approach called High-Discriminative Attribute Feature Learning for Generalized Zero-Shot Learning (HDAFL)
HDAFL utilizes multiple convolutional kernels to automatically learn discriminative regions highly correlated with attributes in images.
We also introduce a Transformer-based attribute discrimination encoder to enhance the discriminative capability among attributes.
arXiv Detail & Related papers (2024-04-07T13:17:47Z) - Emotic Masked Autoencoder with Attention Fusion for Facial Expression Recognition [1.4374467687356276]
This paper presents an innovative approach integrating the MAE-Face self-supervised learning (SSL) method and multi-view Fusion Attention mechanism for expression classification.
We suggest easy-to-implement and no-training frameworks aimed at highlighting key facial features to determine if such features can serve as guides for the model.
The efficacy of this method is validated by improvements in model performance on the Aff-wild2 dataset.
arXiv Detail & Related papers (2024-03-19T16:21:47Z) - Multimodal Adaptive Fusion of Face and Gait Features using Keyless
attention based Deep Neural Networks for Human Identification [67.64124512185087]
Soft biometrics such as gait are widely used with face in surveillance tasks like person recognition and re-identification.
We propose a novel adaptive multi-biometric fusion strategy for the dynamic incorporation of gait and face biometric cues by leveraging keyless attention deep neural networks.
arXiv Detail & Related papers (2023-03-24T05:28:35Z) - CIAO! A Contrastive Adaptation Mechanism for Non-Universal Facial
Expression Recognition [80.07590100872548]
We propose Contrastive Inhibitory Adaptati On (CIAO), a mechanism that adapts the last layer of facial encoders to depict specific affective characteristics on different datasets.
CIAO presents an improvement in facial expression recognition performance over six different datasets with very unique affective representations.
arXiv Detail & Related papers (2022-08-10T15:46:05Z) - TransFA: Transformer-based Representation for Face Attribute Evaluation [87.09529826340304]
We propose a novel textbftransformer-based representation for textbfattribute evaluation method (textbfTransFA)
The proposed TransFA achieves superior performances compared with state-of-the-art methods.
arXiv Detail & Related papers (2022-07-12T10:58:06Z) - Classifying Eye-Tracking Data Using Saliency Maps [8.524684315458245]
This paper proposes a visual saliency based novel feature extraction method for automatic and quantitative classification of eye-tracking data.
Comparing the saliency amplitudes, similarity and dissimilarity of saliency maps with the corresponding eye fixations maps gives an extra dimension of information which is effectively utilized to generate discriminative features to classify the eye-tracking data.
arXiv Detail & Related papers (2020-10-24T15:18:07Z) - View-Invariant Gait Recognition with Attentive Recurrent Learning of
Partial Representations [27.33579145744285]
We propose a network that first learns to extract gait convolutional energy maps (GCEM) from frame-level convolutional features.
It then adopts a bidirectional neural network to learn from split bins of the GCEM, thus exploiting the relations between learned partial recurrent representations.
Our proposed model has been extensively tested on two large-scale CASIA-B and OU-M gait datasets.
arXiv Detail & Related papers (2020-10-18T20:20:43Z) - Dual-Attention GAN for Large-Pose Face Frontalization [59.689836951934694]
We present a novel Dual-Attention Generative Adversarial Network (DA-GAN) for photo-realistic face frontalization.
Specifically, a self-attention-based generator is introduced to integrate local features with their long-range dependencies.
A novel face-attention-based discriminator is applied to emphasize local features of face regions.
arXiv Detail & Related papers (2020-02-17T20:00:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.