Efficient Cloud-edge Collaborative Inference for Object
Re-identification
- URL: http://arxiv.org/abs/2401.02041v1
- Date: Thu, 4 Jan 2024 02:56:50 GMT
- Title: Efficient Cloud-edge Collaborative Inference for Object
Re-identification
- Authors: Chuanming Wang, Yuxin Yang, Mengshi Qi, Huadong Ma
- Abstract summary: We pioneer a cloud-edge collaborative inference framework for ReID systems.
We propose a distribution-aware correlation modeling network (DaCM) to make the desired image return to the cloud server.
DaCM embeds the spatial-temporal correlations implicitly included in the timestamps into a graph structure, and it can be applied in the cloud to regulate the size of the upload window.
- Score: 27.952445808987036
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Current object re-identification (ReID) system follows the centralized
processing paradigm, i.e., all computations are conducted in the cloud server
and edge devices are only used to capture and send images. As the number of
videos experiences a rapid escalation, this paradigm has become impractical due
to the finite computational resources. In such a scenario, the ReID system
should be converted to fit in the cloud-edge collaborative processing paradigm,
which is crucial to boost the scalability and practicality of ReID systems.
However, current relevant work lacks research on this issue, making it
challenging for ReID methods to be adapted effectively. Therefore, we pioneer a
cloud-edge collaborative inference framework for ReID systems and particularly
propose a distribution-aware correlation modeling network (DaCM) to make the
desired image return to the cloud server as soon as possible via learning to
model the spatial-temporal correlations among instances. DaCM embeds the
spatial-temporal correlations implicitly included in the timestamps into a
graph structure, and it can be applied in the cloud to regulate the size of the
upload window and on the edge device to adjust the sequence of images,
respectively. Traditional ReID methods can be combined with DaCM seamlessly,
enabling their application within our proposed edge-cloud collaborative
framework. Extensive experiments demonstrate that our method obviously reduces
transmission overhead and significantly improves performance. We will release
our code and model.
Related papers
- Rendering-Oriented 3D Point Cloud Attribute Compression using Sparse Tensor-based Transformer [52.40992954884257]
3D visualization techniques have fundamentally transformed how we interact with digital content.
Massive data size of point clouds presents significant challenges in data compression.
We propose an end-to-end deep learning framework that seamlessly integrates PCAC with differentiable rendering.
arXiv Detail & Related papers (2024-11-12T16:12:51Z) - Task-Oriented Real-time Visual Inference for IoVT Systems: A Co-design Framework of Neural Networks and Edge Deployment [61.20689382879937]
Task-oriented edge computing addresses this by shifting data analysis to the edge.
Existing methods struggle to balance high model performance with low resource consumption.
We propose a novel co-design framework to optimize neural network architecture.
arXiv Detail & Related papers (2024-10-29T19:02:54Z) - CMR-Agent: Learning a Cross-Modal Agent for Iterative Image-to-Point Cloud Registration [2.400446821380503]
Image-to-point cloud registration aims to determine the relative camera pose of an RGB image with respect to a point cloud.
Most learning-based methods establish 2D-3D point correspondences in feature space without any feedback mechanism for iterative optimization.
We propose to reformulate the registration procedure as an iterative Markov decision process, allowing for incremental adjustments to the camera pose.
arXiv Detail & Related papers (2024-08-05T11:40:59Z) - VICAN: Very Efficient Calibration Algorithm for Large Camera Networks [49.17165360280794]
We introduce a novel methodology that extends Pose Graph Optimization techniques.
We consider the bipartite graph encompassing cameras, object poses evolving dynamically, and camera-object relative transformations at each time step.
Our framework retains compatibility with traditional PGO solvers, but its efficacy benefits from a custom-tailored optimization scheme.
arXiv Detail & Related papers (2024-03-25T17:47:03Z) - RecNet: An Invertible Point Cloud Encoding through Range Image
Embeddings for Multi-Robot Map Sharing and Reconstruction [8.602553195689513]
RecNet is a novel approach to effective place recognition in resource-constrained robots.
It projects 3D point clouds into range images, compresses them using an encoder-decoder framework, and subsequently reconstructs the range image, restoring the original point cloud.
Our proposed approach is assessed using both a publicly available dataset and field experiments$1$, confirming its efficacy and potential for real-world applications.
arXiv Detail & Related papers (2024-02-03T15:43:50Z) - Unified-Width Adaptive Dynamic Network for All-In-One Image Restoration [50.81374327480445]
We introduce a novel concept positing that intricate image degradation can be represented in terms of elementary degradation.
We propose the Unified-Width Adaptive Dynamic Network (U-WADN), consisting of two pivotal components: a Width Adaptive Backbone (WAB) and a Width Selector (WS)
The proposed U-WADN achieves better performance while simultaneously reducing up to 32.3% of FLOPs and providing approximately 15.7% real-time acceleration.
arXiv Detail & Related papers (2024-01-24T04:25:12Z) - GP-PCS: One-shot Feature-Preserving Point Cloud Simplification with Gaussian Processes on Riemannian Manifolds [2.8811433060309763]
We propose a novel, one-shot point cloud simplification method.
It preserves both the salient structural features and the overall shape of a point cloud without any prior surface reconstruction step.
We evaluate our method on several benchmark and self-acquired point clouds, compare it to a range of existing methods, demonstrate its application in downstream tasks of registration and surface reconstruction.
arXiv Detail & Related papers (2023-03-27T14:05:34Z) - Contrastive Learning for Compact Single Image Dehazing [41.83007400559068]
We propose a novel contrastive regularization (CR) built upon contrastive learning to exploit both the information of hazy images and clear images as negative and positive samples.
CR ensures that the restored image is pulled to closer to the clear image and pushed to far away from the hazy image in the representation space.
Considering trade-off between performance and memory storage, we develop a compact dehazing network based on autoencoder-like framework.
arXiv Detail & Related papers (2021-04-19T14:56:21Z) - Retrieve Fast, Rerank Smart: Cooperative and Joint Approaches for
Improved Cross-Modal Retrieval [80.35589927511667]
Current state-of-the-art approaches to cross-modal retrieval process text and visual input jointly, relying on Transformer-based architectures with cross-attention mechanisms that attend over all words and objects in an image.
We propose a novel fine-tuning framework which turns any pretrained text-image multi-modal model into an efficient retrieval model.
Our experiments on a series of standard cross-modal retrieval benchmarks in monolingual, multilingual, and zero-shot setups, demonstrate improved accuracy and huge efficiency benefits over the state-of-the-art cross-encoders.
arXiv Detail & Related papers (2021-03-22T15:08:06Z) - Identity-Aware Attribute Recognition via Real-Time Distributed Inference
in Mobile Edge Clouds [53.07042574352251]
We design novel models for pedestrian attribute recognition with re-ID in an MEC-enabled camera monitoring system.
We propose a novel inference framework with a set of distributed modules, by jointly considering the attribute recognition and person re-ID.
We then devise a learning-based algorithm for the distributions of the modules of the proposed distributed inference framework.
arXiv Detail & Related papers (2020-08-12T12:03:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.