Scalable Human-Machine Point Cloud Compression
- URL: http://arxiv.org/abs/2402.12532v3
- Date: Fri, 23 Feb 2024 18:41:15 GMT
- Title: Scalable Human-Machine Point Cloud Compression
- Authors: Mateen Ulhaq, Ivan V. Baji\'c
- Abstract summary: In this paper, we present a scalable for point-cloud data that is specialized for the machine task of classification, while also providing a mechanism for human viewing.
In the proposed scalable, the "base" bitstream supports the machine task, and an "enhancement" bitstream may be used for better input reconstruction performance for human viewing.
- Score: 29.044369073873465
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Due to the limited computational capabilities of edge devices, deep learning
inference can be quite expensive. One remedy is to compress and transmit point
cloud data over the network for server-side processing. Unfortunately, this
approach can be sensitive to network factors, including available bitrate.
Luckily, the bitrate requirements can be reduced without sacrificing inference
accuracy by using a machine task-specialized codec. In this paper, we present a
scalable codec for point-cloud data that is specialized for the machine task of
classification, while also providing a mechanism for human viewing. In the
proposed scalable codec, the "base" bitstream supports the machine task, and an
"enhancement" bitstream may be used for better input reconstruction performance
for human viewing. We base our architecture on PointNet++, and test its
efficacy on the ModelNet40 dataset. We show significant improvements over prior
non-specialized codecs.
Related papers
- Plug-and-Play Versatile Compressed Video Enhancement [57.62582951699999]
Video compression effectively reduces the size of files, making it possible for real-time cloud computing.
However, it comes at the cost of visual quality, challenges the robustness of downstream vision models.
We present a versatile-aware enhancement framework that adaptively enhance videos under different compression settings.
arXiv Detail & Related papers (2025-04-21T18:39:31Z) - Embedding Compression Distortion in Video Coding for Machines [67.97469042910855]
Currently, video transmission serves not only the Human Visual System (HVS) for viewing but also machine perception for analysis.
We propose a Compression Distortion Embedding (CDRE) framework, which extracts machine-perception-related distortion representation and embeds it into downstream models.
Our framework can effectively boost the rate-task performance of existing codecs with minimal overhead in terms of execution time, and number of parameters.
arXiv Detail & Related papers (2025-03-27T13:01:53Z) - RL-RC-DoT: A Block-level RL agent for Task-Aware Video Compression [68.31184784672227]
In modern applications such as autonomous driving, an overwhelming majority of videos serve as input for AI systems performing tasks.
It is therefore useful to optimize the encoder for a downstream task instead of for image quality.
Here, we address this challenge by controlling the Quantization Parameters (QPs) at the macro-block level to optimize the downstream task.
arXiv Detail & Related papers (2025-01-21T15:36:08Z) - Towards Point Cloud Compression for Machine Perception: A Simple and Strong Baseline by Learning the Octree Depth Level Predictor [12.510990055381452]
We propose a point cloud compression framework that simultaneously handles both human and machine vision tasks.
Our framework learns a scalable bit-stream, using only subsets for different machine vision tasks to save bit-rate.
A new octree depth-level predictor adaptively determines the optimal depth level for each octree constructed from a point cloud.
arXiv Detail & Related papers (2024-06-02T16:13:57Z) - A Perspective on Deep Vision Performance with Standard Image and Video Codecs [41.73262031925552]
Resource-constrained hardware, such as edge devices or cell phones, often rely on cloud servers to provide the required computational resources for inference in deep vision models.
This paper aims to examine the implications of employing standardized codecs within deep vision pipelines.
We find that using JPEG and H.264 coding significantly deteriorates the accuracy across a broad range of vision tasks and models.
arXiv Detail & Related papers (2024-04-18T16:58:05Z) - Learned Point Cloud Compression for Classification [35.103437828235826]
Deep learning is increasingly being used to perform machine vision tasks such as classification, object detection, and segmentation on 3D point cloud data.
We present a novel point cloud that is highly specialized for the machine task of classification.
In particular, it achieves a 93% reduction in BD-bitrate over non-specialized codecs on the ModelNet40 dataset.
arXiv Detail & Related papers (2023-08-11T06:28:19Z) - Attention-based Feature Compression for CNN Inference Offloading in Edge
Computing [93.67044879636093]
This paper studies the computational offloading of CNN inference in device-edge co-inference systems.
We propose a novel autoencoder-based CNN architecture (AECNN) for effective feature extraction at end-device.
Experiments show that AECNN can compress the intermediate data by more than 256x with only about 4% accuracy loss.
arXiv Detail & Related papers (2022-11-24T18:10:01Z) - Preprocessing Enhanced Image Compression for Machine Vision [14.895698385236937]
We propose a preprocessing enhanced image compression method for machine vision tasks.
Our framework is built upon the traditional non-differential codecs.
Experimental results show our method achieves a better tradeoff between the coding and the performance of the downstream machine vision tasks by saving about 20%.
arXiv Detail & Related papers (2022-06-12T03:36:38Z) - SoftPool++: An Encoder-Decoder Network for Point Cloud Completion [93.54286830844134]
We propose a novel convolutional operator for the task of point cloud completion.
The proposed operator does not require any max-pooling or voxelization operation.
We show that our approach achieves state-of-the-art performance in shape completion at low and high resolutions.
arXiv Detail & Related papers (2022-05-08T15:31:36Z) - A New Image Codec Paradigm for Human and Machine Uses [53.48873918537017]
A new scalable image paradigm for both human and machine uses is proposed in this work.
The high-level instance segmentation map and the low-level signal features are extracted with neural networks.
An image is designed and trained to achieve the general-quality image reconstruction with the 16-bit gray-scale profile and signal features.
arXiv Detail & Related papers (2021-12-19T06:17:38Z) - Small Lesion Segmentation in Brain MRIs with Subpixel Embedding [105.1223735549524]
We present a method to segment MRI scans of the human brain into ischemic stroke lesion and normal tissues.
We propose a neural network architecture in the form of a standard encoder-decoder where predictions are guided by a spatial expansion embedding network.
arXiv Detail & Related papers (2021-09-18T00:21:17Z) - Dynamic Neural Representational Decoders for High-Resolution Semantic
Segmentation [98.05643473345474]
We propose a novel decoder, termed dynamic neural representational decoder (NRD)
As each location on the encoder's output corresponds to a local patch of the semantic labels, in this work, we represent these local patches of labels with compact neural networks.
This neural representation enables our decoder to leverage the smoothness prior in the semantic label space, and thus makes our decoder more efficient.
arXiv Detail & Related papers (2021-07-30T04:50:56Z) - Quantized Neural Networks via {-1, +1} Encoding Decomposition and
Acceleration [83.84684675841167]
We propose a novel encoding scheme using -1, +1 to decompose quantized neural networks (QNNs) into multi-branch binary networks.
We validate the effectiveness of our method on large-scale image classification, object detection, and semantic segmentation tasks.
arXiv Detail & Related papers (2021-06-18T03:11:15Z) - End-to-end optimized image compression for machines, a study [3.0448872422956437]
An increasing share of image and video content is analyzed by machines rather than viewed by humans.
Conventional coding tools are challenging to specialize for machine tasks as they were originally designed for human perception.
neural network based codecs can be jointly trained end-to-end with any convolutional neural network (CNN)-based task model.
arXiv Detail & Related papers (2020-11-10T20:10:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.