DNN-Driven Compressive Offloading for Edge-Assisted Semantic Video
Segmentation
- URL: http://arxiv.org/abs/2203.14481v1
- Date: Mon, 28 Mar 2022 03:49:57 GMT
- Title: DNN-Driven Compressive Offloading for Edge-Assisted Semantic Video
Segmentation
- Authors: Xuedou Xiao, Juecheng Zhang, Wei Wang, Jianhua He, Qian Zhang
- Abstract summary: This paper introduces STAC, a uniform compression scheme tailored for edge-assisted semantic segmentation.
We show that STAC can save up to 20.95% of bandwidth without losing accuracy, compared to the state-of-the-art algorithm.
- Score: 7.9955920576321615
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Deep learning has shown impressive performance in semantic segmentation, but
it is still unaffordable for resource-constrained mobile devices. While
offloading computation tasks is promising, the high traffic demands overwhelm
the limited bandwidth. Existing compression algorithms are not fit for semantic
segmentation, as the lack of obvious and concentrated regions of interest
(RoIs) forces the adoption of uniform compression strategies, leading to low
compression ratios or accuracy. This paper introduces STAC, a DNN-driven
compression scheme tailored for edge-assisted semantic video segmentation. STAC
is the first to exploit DNN's gradients as spatial sensitivity metrics for
spatial adaptive compression and achieves superior compression ratio and
accuracy. Yet, it is challenging to adapt this content-customized compression
to videos. Practical issues include varying spatial sensitivity and huge
bandwidth consumption for compression strategy feedback and offloading. We
tackle these issues through a spatiotemporal adaptive scheme, which (1) takes
partial strategy generation operations offline to reduce communication load,
and (2) propagates compression strategies and segmentation results across
frames through dense optical flow, and adaptively offloads keyframes to
accommodate video content. We implement STAC on a commodity mobile device.
Experiments show that STAC can save up to 20.95% of bandwidth without losing
accuracy, compared to the state-of-the-art algorithm.
Related papers
- LoRC: Low-Rank Compression for LLMs KV Cache with a Progressive Compression Strategy [59.1298692559785]
Key-Value ( KV) cache is crucial component in serving transformer-based autoregressive large language models (LLMs)
Existing approaches to mitigate this issue include: (1) efficient attention variants integrated in upcycling stages; (2) KV cache compression at test time; and (3) KV cache compression at test time.
We propose a low-rank approximation of KV weight matrices, allowing plug-in integration with existing transformer-based LLMs without model retraining.
Our method is designed to function without model tuning in upcycling stages or task-specific profiling in test stages.
arXiv Detail & Related papers (2024-10-04T03:10:53Z) - Accelerating Learned Video Compression via Low-Resolution Representation Learning [18.399027308582596]
We introduce an efficiency-optimized framework for learned video compression that focuses on low-resolution representation learning.
Our method achieves performance levels on par with the low-decay P configuration of the H.266 reference software VTM.
arXiv Detail & Related papers (2024-07-23T12:02:57Z) - Efficient Semantic Segmentation by Altering Resolutions for Compressed
Videos [42.944135041061166]
We propose an altering resolution framework called AR-Seg for compressed videos to achieve efficient video segmentation.
AR-Seg aims to reduce the computational cost by using low resolution for non-keyframes.
Experiments on CamVid and Cityscapes show that AR-Seg achieves state-of-the-art performance.
arXiv Detail & Related papers (2023-03-13T15:58:15Z) - FrankenSplit: Efficient Neural Feature Compression with Shallow Variational Bottleneck Injection for Mobile Edge Computing [5.815300670677979]
We introduce a novel framework for resource-conscious compression models and extensively evaluate our method in an asymmetric environment.
Our method achieves 60% lower than a state-of-the-art SC method without decreasing accuracy and is up 16x faster than offloading with existing standards.
arXiv Detail & Related papers (2023-02-21T14:03:22Z) - Deep Lossy Plus Residual Coding for Lossless and Near-lossless Image
Compression [85.93207826513192]
We propose a unified and powerful deep lossy plus residual (DLPR) coding framework for both lossless and near-lossless image compression.
We solve the joint lossy and residual compression problem in the approach of VAEs.
In the near-lossless mode, we quantize the original residuals to satisfy a given $ell_infty$ error bound.
arXiv Detail & Related papers (2022-09-11T12:11:56Z) - Cross Modal Compression: Towards Human-comprehensible Semantic
Compression [73.89616626853913]
Cross modal compression is a semantic compression framework for visual data.
We show that our proposed CMC can achieve encouraging reconstructed results with an ultrahigh compression ratio.
arXiv Detail & Related papers (2022-09-06T15:31:11Z) - Learned Video Compression via Heterogeneous Deformable Compensation
Network [78.72508633457392]
We propose a learned video compression framework via heterogeneous deformable compensation strategy (HDCVC) to tackle the problems of unstable compression performance.
More specifically, the proposed algorithm extracts features from the two adjacent frames to estimate content-Neighborhood heterogeneous deformable (HetDeform) kernel offsets.
Experimental results indicate that HDCVC achieves superior performance than the recent state-of-the-art learned video compression approaches.
arXiv Detail & Related papers (2022-07-11T02:31:31Z) - Towards Compact CNNs via Collaborative Compression [166.86915086497433]
We propose a Collaborative Compression scheme, which joints channel pruning and tensor decomposition to compress CNN models.
We achieve 52.9% FLOPs reduction by removing 48.4% parameters on ResNet-50 with only a Top-1 accuracy drop of 0.56% on ImageNet 2012.
arXiv Detail & Related papers (2021-05-24T12:07:38Z) - Content Adaptive and Error Propagation Aware Deep Video Compression [110.31693187153084]
We propose a content adaptive and error propagation aware video compression system.
Our method employs a joint training strategy by considering the compression performance of multiple consecutive frames instead of a single frame.
Instead of using the hand-crafted coding modes in the traditional compression systems, we design an online encoder updating scheme in our system.
arXiv Detail & Related papers (2020-03-25T09:04:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.