Just Noticeable Difference Modeling for Face Recognition System
- URL: http://arxiv.org/abs/2209.05856v2
- Date: Thu, 28 Sep 2023 13:29:16 GMT
- Title: Just Noticeable Difference Modeling for Face Recognition System
- Authors: Yu Tian and Zhangkai Ni and Baoliang Chen and Shurun Wang and Shiqi
Wang and Hanli Wang and Sam Kwong
- Abstract summary: We make the first attempt to study just noticeable difference (JND) for the automatic face recognition system.
We develop a novel JND prediction model to directly infer JND images for the FR system.
Experimental results have demonstrated that the proposed model achieves higher accuracy of JND map prediction.
- Score: 69.28990314553076
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: High-quality face images are required to guarantee the stability and
reliability of automatic face recognition (FR) systems in surveillance and
security scenarios. However, a massive amount of face data is usually
compressed before being analyzed due to limitations on transmission or storage.
The compressed images may lose the powerful identity information, resulting in
the performance degradation of the FR system. Herein, we make the first attempt
to study just noticeable difference (JND) for the FR system, which can be
defined as the maximum distortion that the FR system cannot notice. More
specifically, we establish a JND dataset including 3530 original images and
137,670 compressed images generated by advanced reference encoding/decoding
software based on the Versatile Video Coding (VVC) standard (VTM-15.0).
Subsequently, we develop a novel JND prediction model to directly infer JND
images for the FR system. In particular, in order to maximum redundancy removal
without impairment of robust identity information, we apply the encoder with
multiple feature extraction and attention-based feature decomposition modules
to progressively decompose face features into two uncorrelated components,
i.e., identity and residual features, via self-supervised learning. Then, the
residual feature is fed into the decoder to generate the residual map. Finally,
the predicted JND map is obtained by subtracting the residual map from the
original image. Experimental results have demonstrated that the proposed model
achieves higher accuracy of JND map prediction compared with the
state-of-the-art JND models, and is capable of saving more bits while
maintaining the performance of the FR system compared with VTM-15.0.
Related papers
- High-Precision Dichotomous Image Segmentation via Probing Diffusion Capacity [69.32473738284374]
We propose DiffDIS, a diffusion-driven segmentation model that taps into the potential of the pre-trained U-Net within diffusion models.
By leveraging the robust generalization capabilities and rich, versatile image representation prior to the SD models, we significantly reduce the inference time while preserving high-fidelity, detailed generation.
Experiments on the DIS5K dataset demonstrate the superiority of DiffDIS, achieving state-of-the-art results through a streamlined inference process.
arXiv Detail & Related papers (2024-10-14T02:49:23Z) - Multi-Feature Aggregation in Diffusion Models for Enhanced Face Super-Resolution [6.055006354743854]
We develop an algorithm that utilize a low-resolution image combined with features extracted from multiple low-quality images to generate a super-resolved image.
Unlike other algorithms, our approach recovers facial features without explicitly providing attribute information.
This is the first time multi-features combined with low-resolution images are used as conditioners to generate more reliable super-resolution images.
arXiv Detail & Related papers (2024-08-27T20:08:33Z) - SG-JND: Semantic-Guided Just Noticeable Distortion Predictor For Image Compression [50.2496399381438]
Just noticeable distortion (JND) represents the threshold of distortion in an image that is minimally perceptible to the human visual system.
Traditional JND prediction methods only rely on pixel-level or sub-band level features.
We propose a Semantic-Guided JND network to leverage semantic information for JND prediction.
arXiv Detail & Related papers (2024-08-08T07:14:57Z) - NAF-DPM: A Nonlinear Activation-Free Diffusion Probabilistic Model for Document Enhancement [4.841365627573421]
A crucial preprocessing step is essential to eliminate noise while preserving text and key features of documents.
We propose NAF-DPM, a novel generative framework based on a diffusion probabilistic model (DPM) designed to restore the original quality of degraded documents.
arXiv Detail & Related papers (2024-04-08T16:52:21Z) - DiAD: A Diffusion-based Framework for Multi-class Anomaly Detection [55.48770333927732]
We propose a Difusion-based Anomaly Detection (DiAD) framework for multi-class anomaly detection.
It consists of a pixel-space autoencoder, a latent-space Semantic-Guided (SG) network with a connection to the stable diffusion's denoising network, and a feature-space pre-trained feature extractor.
Experiments on MVTec-AD and VisA datasets demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2023-12-11T18:38:28Z) - Restoration of the JPEG Maximum Lossy Compressed Face Images with
Hourglass Block based on Early Stopping Discriminator [0.0]
This paper addresses the restoration of JPEG images that have suffered significant loss due to maximum compression using a GAN-based net-work method.
The network incorporates two loss functions, LF Loss and HF Loss, to generate natural and high-performance images.
Results show that the blocking phe-nomenon in lost compressed images was removed, and recognizable identities were generated.
arXiv Detail & Related papers (2023-06-22T09:21:48Z) - Multi-Prior Learning via Neural Architecture Search for Blind Face
Restoration [61.27907052910136]
Blind Face Restoration (BFR) aims to recover high-quality face images from low-quality ones.
Current methods still suffer from two major difficulties: 1) how to derive a powerful network architecture without extensive hand tuning; 2) how to capture complementary information from multiple facial priors in one network to improve restoration performance.
We propose a Face Restoration Searching Network (FRSNet) to adaptively search the suitable feature extraction architecture within our specified search space.
arXiv Detail & Related papers (2022-06-28T12:29:53Z) - STIP: A SpatioTemporal Information-Preserving and Perception-Augmented
Model for High-Resolution Video Prediction [78.129039340528]
We propose a Stemporal Information-Preserving and Perception-Augmented Model (STIP) to solve the above two problems.
The proposed model aims to preserve thetemporal information for videos during the feature extraction and the state transitions.
Experimental results show that the proposed STIP can predict videos with more satisfactory visual quality compared with a variety of state-of-the-art methods.
arXiv Detail & Related papers (2022-06-09T09:49:04Z) - One-Stage Deep Edge Detection Based on Dense-Scale Feature Fusion and
Pixel-Level Imbalance Learning [5.370848116287344]
We propose a one-stage neural network model that can generate high-quality edge images without postprocessing.
The proposed model adopts a classic encoder-decoder framework in which a pre-trained neural model is used as the encoder.
We propose a new loss function that addresses the pixel-level imbalance in the edge image.
arXiv Detail & Related papers (2022-03-17T15:26:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.