End-to-End Facial Deep Learning Feature Compression with Teacher-Student
Enhancement
- URL: http://arxiv.org/abs/2002.03627v1
- Date: Mon, 10 Feb 2020 10:08:44 GMT
- Title: End-to-End Facial Deep Learning Feature Compression with Teacher-Student
Enhancement
- Authors: Shurun Wang, Wenhan Yang, Shiqi Wang
- Abstract summary: We propose a novel end-to-end feature compression scheme by leveraging the representation and learning capability of deep neural networks.
In particular, the extracted features are compactly coded in an end-to-end manner by optimizing the rate-distortion cost.
We verify the effectiveness of the proposed model with the facial feature, and experimental results reveal better compression performance in terms of rate-accuracy.
- Score: 57.18801093608717
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we propose a novel end-to-end feature compression scheme by
leveraging the representation and learning capability of deep neural networks,
towards intelligent front-end equipped analysis with promising accuracy and
efficiency. In particular, the extracted features are compactly coded in an
end-to-end manner by optimizing the rate-distortion cost to achieve
feature-in-feature representation. In order to further improve the compression
performance, we present a latent code level teacher-student enhancement model,
which could efficiently transfer the low bit-rate representation into a high
bit rate one. Such a strategy further allows us to adaptively shift the
representation cost to decoding computations, leading to more flexible feature
compression with enhanced decoding capability. We verify the effectiveness of
the proposed model with the facial feature, and experimental results reveal
better compression performance in terms of rate-accuracy compared with existing
models.
Related papers
- Multi-Scale Feature Prediction with Auxiliary-Info for Neural Image Compression [13.076563599765176]
We introduce a new predictive structure consisting of the auxiliary coarse network and the main network, inspired by neural video compression.
Our model outperforms other neural image compression models and achieves a 19.49% higher rate-distortion performance than VVC on Tecnick dataset.
arXiv Detail & Related papers (2024-09-19T12:41:53Z) - Joint Hierarchical Priors and Adaptive Spatial Resolution for Efficient
Neural Image Compression [11.25130799452367]
We propose an absolute image compression transformer (ICT) for neural image compression (NIC)
ICT captures both global and local contexts from the latent representations and better parameterize the distribution of the quantized latents.
Our framework significantly improves the trade-off between coding efficiency and decoder complexity over the versatile video coding (VVC) reference encoder (VTM-18.0) and the neural SwinT-ChARM.
arXiv Detail & Related papers (2023-07-05T13:17:14Z) - Learning Accurate Performance Predictors for Ultrafast Automated Model
Compression [86.22294249097203]
We propose an ultrafast automated model compression framework called SeerNet for flexible network deployment.
Our method achieves competitive accuracy-complexity trade-offs with significant reduction of the search cost.
arXiv Detail & Related papers (2023-04-13T10:52:49Z) - LVQAC: Lattice Vector Quantization Coupled with Spatially Adaptive
Companding for Efficient Learned Image Compression [24.812267280543693]
We present a novel Lattice Vector Quantization scheme coupled with a spatially Adaptive Companding (LVQAC) mapping.
For any end-to-end CNN image compression models, replacing uniform quantizer by LVQAC achieves better rate-distortion performance without significantly increasing the model complexity.
arXiv Detail & Related papers (2023-03-25T23:34:15Z) - Estimating the Resize Parameter in End-to-end Learned Image Compression [50.20567320015102]
We describe a search-free resizing framework that can further improve the rate-distortion tradeoff of recent learned image compression models.
Our results show that our new resizing parameter estimation framework can provide Bjontegaard-Delta rate (BD-rate) improvement of about 10% against leading perceptual quality engines.
arXiv Detail & Related papers (2022-04-26T01:35:02Z) - Video Coding for Machine: Compact Visual Representation Compression for
Intelligent Collaborative Analytics [101.35754364753409]
Video Coding for Machines (VCM) is committed to bridging to an extent separate research tracks of video/image compression and feature compression.
This paper summarizes VCM methodology and philosophy based on existing academia and industrial efforts.
arXiv Detail & Related papers (2021-10-18T12:42:13Z) - End-to-end Compression Towards Machine Vision: Network Architecture
Design and Optimization [21.64194147628761]
We show that the design and optimization of network architecture could be further improved for compression towards machine vision.
We propose an inverted bottleneck structure for end-to-end compression towards machine vision.
We show that the proposed scheme achieves significant BD-rate savings in terms of analysis performance.
arXiv Detail & Related papers (2021-07-01T09:36:32Z) - Towards Analysis-friendly Face Representation with Scalable Feature and
Texture Compression [113.30411004622508]
We show that a universal and collaborative visual information representation can be achieved in a hierarchical way.
Based on the strong generative capability of deep neural networks, the gap between the base feature layer and enhancement layer is further filled with the feature level texture reconstruction.
To improve the efficiency of the proposed framework, the base layer neural network is trained in a multi-task manner.
arXiv Detail & Related papers (2020-04-21T14:32:49Z) - Structured Sparsification with Joint Optimization of Group Convolution
and Channel Shuffle [117.95823660228537]
We propose a novel structured sparsification method for efficient network compression.
The proposed method automatically induces structured sparsity on the convolutional weights.
We also address the problem of inter-group communication with a learnable channel shuffle mechanism.
arXiv Detail & Related papers (2020-02-19T12:03:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.