Towards Analysis-friendly Face Representation with Scalable Feature and
Texture Compression
- URL: http://arxiv.org/abs/2004.10043v2
- Date: Mon, 19 Apr 2021 16:40:09 GMT
- Title: Towards Analysis-friendly Face Representation with Scalable Feature and
Texture Compression
- Authors: Shurun Wang, Shiqi Wang, Wenhan Yang, Xinfeng Zhang, Shanshe Wang,
Siwei Ma, Wen Gao
- Abstract summary: We show that a universal and collaborative visual information representation can be achieved in a hierarchical way.
Based on the strong generative capability of deep neural networks, the gap between the base feature layer and enhancement layer is further filled with the feature level texture reconstruction.
To improve the efficiency of the proposed framework, the base layer neural network is trained in a multi-task manner.
- Score: 113.30411004622508
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: It plays a fundamental role to compactly represent the visual information
towards the optimization of the ultimate utility in myriad visual data centered
applications. With numerous approaches proposed to efficiently compress the
texture and visual features serving human visual perception and machine
intelligence respectively, much less work has been dedicated to studying the
interactions between them. Here we investigate the integration of feature and
texture compression, and show that a universal and collaborative visual
information representation can be achieved in a hierarchical way. In
particular, we study the feature and texture compression in a scalable coding
framework, where the base layer serves as the deep learning feature and
enhancement layer targets to perfectly reconstruct the texture. Based on the
strong generative capability of deep neural networks, the gap between the base
feature layer and enhancement layer is further filled with the feature level
texture reconstruction, aiming to further construct texture representation from
feature. As such, the residuals between the original and reconstructed texture
could be further conveyed in the enhancement layer. To improve the efficiency
of the proposed framework, the base layer neural network is trained in a
multi-task manner such that the learned features enjoy both high quality
reconstruction and high accuracy analysis. We further demonstrate the framework
and optimization strategies in face image compression, and promising coding
performance has been achieved in terms of both rate-fidelity and rate-accuracy.
Related papers
- Toward Scalable Image Feature Compression: A Content-Adaptive and Diffusion-Based Approach [44.03561901593423]
This paper introduces a content-adaptive diffusion model for scalable image compression.
The proposed method encodes fine textures through a diffusion process, enhancing perceptual quality.
Experiments demonstrate the effectiveness of the proposed framework in both image reconstruction and downstream machine vision tasks.
arXiv Detail & Related papers (2024-10-08T15:48:34Z) - Texture-guided Coding for Deep Features [33.05814372247946]
This paper investigates features and textures and proposes a texture-guided feature compression strategy based on their characteristics.
The strategy comprises feature layers and texture layers. The feature layers serve the machine, including a feature selection module and a feature reconstruction network.
With the assistance of texture images, they selectively compress and transmit channels relevant to visual tasks, reducing feature data while providing high-quality features for the machine.
Our method fully exploits the characteristics of texture and features. It eliminates feature redundancy, reconstructs high-quality preview images for humans, and supports decision-making.
arXiv Detail & Related papers (2024-05-30T03:38:44Z) - ENTED: Enhanced Neural Texture Extraction and Distribution for
Reference-based Blind Face Restoration [51.205673783866146]
We present ENTED, a new framework for blind face restoration that aims to restore high-quality and realistic portrait images.
We utilize a texture extraction and distribution framework to transfer high-quality texture features between the degraded input and reference image.
The StyleGAN-like architecture in our framework requires high-quality latent codes to generate realistic images.
arXiv Detail & Related papers (2024-01-13T04:54:59Z) - Semantic-aware Texture-Structure Feature Collaboration for Underwater
Image Enhancement [58.075720488942125]
Underwater image enhancement has become an attractive topic as a significant technology in marine engineering and aquatic robotics.
We develop an efficient and compact enhancement network in collaboration with a high-level semantic-aware pretrained model.
We also apply the proposed algorithm to the underwater salient object detection task to reveal the favorable semantic-aware ability for high-level vision tasks.
arXiv Detail & Related papers (2022-11-19T07:50:34Z) - Video Coding for Machine: Compact Visual Representation Compression for
Intelligent Collaborative Analytics [101.35754364753409]
Video Coding for Machines (VCM) is committed to bridging to an extent separate research tracks of video/image compression and feature compression.
This paper summarizes VCM methodology and philosophy based on existing academia and industrial efforts.
arXiv Detail & Related papers (2021-10-18T12:42:13Z) - Conceptual Compression via Deep Structure and Texture Synthesis [42.68994438290913]
We propose a novel conceptual compression framework that encodes visual data into compact structure and texture representations, then decodes in a deep synthesis fashion.
In particular, we propose to compress images by a dual-layered model consisting of two complementary visual features.
At the encoder side, the structural maps and texture representations are individually extracted and compressed, generating the compact, interpretable, inter-operable bitstreams.
During the decoding stage, a hierarchical fusion GAN (HF-GAN) is proposed to learn the synthesis paradigm where the textures are rendered into the decoded structural maps, leading to high-quality reconstruction
arXiv Detail & Related papers (2020-11-10T08:48:32Z) - Region-adaptive Texture Enhancement for Detailed Person Image Synthesis [86.69934638569815]
RATE-Net is a novel framework for synthesizing person images with sharp texture details.
The proposed framework leverages an additional texture enhancing module to extract appearance information from the source image.
Experiments conducted on DeepFashion benchmark dataset have demonstrated the superiority of our framework compared with existing networks.
arXiv Detail & Related papers (2020-05-26T02:33:21Z) - End-to-End Facial Deep Learning Feature Compression with Teacher-Student
Enhancement [57.18801093608717]
We propose a novel end-to-end feature compression scheme by leveraging the representation and learning capability of deep neural networks.
In particular, the extracted features are compactly coded in an end-to-end manner by optimizing the rate-distortion cost.
We verify the effectiveness of the proposed model with the facial feature, and experimental results reveal better compression performance in terms of rate-accuracy.
arXiv Detail & Related papers (2020-02-10T10:08:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.