Learning to Localize Through Compressed Binary Maps
- URL: http://arxiv.org/abs/2012.10942v1
- Date: Sun, 20 Dec 2020 14:47:15 GMT
- Title: Learning to Localize Through Compressed Binary Maps
- Authors: Xinkai Wei, Ioan Andrei B\^arsan, Shenlong Wang, Julieta Martinez,
Raquel Urtasun
- Abstract summary: We learn to compress the map representation such that it is optimal for the localization task.
Our experiments show that it is possible to learn a task-specific compression which reduces storage requirements by two orders of magnitude over general-purpose codecs.
- Score: 83.03367511221437
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: One of the main difficulties of scaling current localization systems to large
environments is the on-board storage required for the maps. In this paper we
propose to learn to compress the map representation such that it is optimal for
the localization task. As a consequence, higher compression rates can be
achieved without loss of localization accuracy when compared to standard coding
schemes that optimize for reconstruction, thus ignoring the end task. Our
experiments show that it is possible to learn a task-specific compression which
reduces storage requirements by two orders of magnitude over general-purpose
codecs such as WebP without sacrificing performance.
Related papers
- SSR: A Generic Framework for Text-Aided Map Compression for Localization [13.691397425850097]
We propose a text-enhanced compression framework that reduces both memory and bandwidth footprints while retaining high-fidelity localization.<n>Similarity Space Replication learns an adaptive image embedding in one shot that captures only the information "complementary" to the text descriptions.<n>We validate our compression framework on multiple downstream localization tasks, including Visual Place Recognition and object-centric Monte Carlo localization.
arXiv Detail & Related papers (2026-03-04T16:55:48Z) - CLIP-Map: Structured Matrix Mapping for Parameter-Efficient CLIP Compression [70.45437536012015]
Contrastive Language-Image Pre-training (CLIP) has achieved widely applications in computer vision tasks.<n>CLIP suffers from high memory and computation cost, which prohibits its usage to the resource-limited application scenarios.<n>We propose a novel mapping-based CLIP compression framework, CLIP-Map.
arXiv Detail & Related papers (2026-02-05T17:25:16Z) - Saliency Driven Imagery Preprocessing for Efficient Compression -- Industrial Paper [5.980556660370507]
We show how preprocessing techniques driven by saliency maps can be used to create variable rate image compression within a single large satellite image.<n>Specifically, we use variable sized smoothing kernels that map to different quantized saliency levels to process imagery pixels in order to optimize downstream compression and encoding schemes.
arXiv Detail & Related papers (2026-01-24T18:52:19Z) - ImLoc: Revisiting Visual Localization with Image-based Representation [61.282162006394934]
We propose to augment each image with estimated depth maps to capture the geometric structure.<n>This representation is easy to build and maintain, but achieves highest accuracy in challenging conditions.<n>Our method achieves a new state-of-the-art accuracy on various standard benchmarks and outperforms existing memory-efficient methods at comparable map sizes.
arXiv Detail & Related papers (2026-01-07T18:51:51Z) - Application of Deep Learning in Biological Data Compression [0.0]
Cryogenic electron microscopy (Cryo-EM) has become an essential tool for capturing high-resolution biological structures.<n>Despite its advantage in visualizations, the large storage size of Cryo-EM data file poses significant challenges for researchers and educators.<n>This paper investigates the application of deep learning, specifically implicit neural representation (INR), to compress Cryo-EM biological data.
arXiv Detail & Related papers (2025-12-15T04:40:23Z) - UniGist: Towards General and Hardware-aligned Sequence-level Long Context Compression [86.33995240043936]
UniGist is a sequence-level long-context compression framework for large language models.<n>It efficiently preserves context information by replacing raw tokens with special compression tokens (gists) in a fine-grained manner.<n>Our scheme also supports flexible inference by allowing the actual removal of compressed tokens, resulting in real-time memory savings.
arXiv Detail & Related papers (2025-09-19T08:47:37Z) - Re-Densification Meets Cross-Scale Propagation: Real-Time Neural Compression of LiDAR Point Clouds [83.39320394656855]
LiDAR point clouds are fundamental to various applications, yet high-precision scans incur substantial storage and transmission overhead.<n>Existing methods typically convert unordered points into hierarchical octree or voxel structures for dense-to-sparse predictive coding.<n>Our framework comprises two lightweight modules. First, the Geometry Re-Densification Module re-densifies encoded sparse geometry, extracts features at denser scale, and then re-sparsifies the features for predictive coding.
arXiv Detail & Related papers (2025-08-28T06:36:10Z) - Mustafar: Promoting Unstructured Sparsity for KV Cache Pruning in LLM Inference [2.0449242727404235]
unstructured sparsity significantly improves KV cache compression for LLMs.<n>We find per-token magnitude-based pruning as highly effective for both Key and Value caches under unstructured sparsity.
arXiv Detail & Related papers (2025-05-28T22:32:15Z) - Lossless Compression for LLM Tensor Incremental Snapshots [0.0]
We build an effective compression solution, known as Language Model (LMC)<n>We show that a 16-core parallel implementation of LMC can attain compression and decompression throughput of 2.78 GiB/s and 3.76 GiB/s respectively.<n>This increase in performance ultimately reduces the resources needed and provides more time to copy the data to the storage system before the next epoch thus allowing for higher-frequency checkpoints.
arXiv Detail & Related papers (2025-05-14T21:24:14Z) - Communication-aware Hierarchical Map Compression of Time-Varying Environments for Mobile Robots [2.8544822698499255]
We develop a framework for the time-sequential compression of dynamic probabilistic occupancy grids.
We search for a multi-resolution hierarchical encoder that balances the quality of the compressed map with its description size.
arXiv Detail & Related papers (2025-04-14T22:54:29Z) - DeepFGS: Fine-Grained Scalable Coding for Learned Image Compression [27.834491128701963]
This paper proposes a learned fine-grained scalable image compression framework, namely DeepFGS.
For entropy coding, we design a mutual entropy model to fully explore the correlation between the basic and scalable features.
Experiments demonstrate that our proposed DeepFGS outperforms previous learning-based scalable image compression models.
arXiv Detail & Related papers (2024-11-30T11:19:38Z) - Large Language Models for Lossless Image Compression: Next-Pixel Prediction in Language Space is All You Need [53.584140947828004]
Language large model (LLM) with unprecedented intelligence is a general-purpose lossless compressor for various data modalities.
We propose P$2$-LLM, a next-pixel prediction-based LLM, which integrates various elaborated insights and methodologies.
Experiments on benchmark datasets demonstrate that P$2$-LLM can beat SOTA classical and learned codecs.
arXiv Detail & Related papers (2024-11-19T12:15:40Z) - LoRC: Low-Rank Compression for LLMs KV Cache with a Progressive Compression Strategy [59.1298692559785]
Key-Value ( KV) cache is crucial component in serving transformer-based autoregressive large language models (LLMs)
Existing approaches to mitigate this issue include: (1) efficient attention variants integrated in upcycling stages; (2) KV cache compression at test time; and (3) KV cache compression at test time.
We propose a low-rank approximation of KV weight matrices, allowing plug-in integration with existing transformer-based LLMs without model retraining.
Our method is designed to function without model tuning in upcycling stages or task-specific profiling in test stages.
arXiv Detail & Related papers (2024-10-04T03:10:53Z) - Fast Point Cloud Geometry Compression with Context-based Residual Coding and INR-based Refinement [19.575833741231953]
We use the KNN method to determine the neighborhoods of raw surface points.
A conditional probability model is adaptive to local geometry, leading to significant rate reduction.
We incorporate an implicit neural representation into the refinement layer, allowing the decoder to sample points on the underlying surface at arbitrary densities.
arXiv Detail & Related papers (2024-08-06T05:24:06Z) - Lossless Image Compression Using Multi-level Dictionaries: Binary Images [2.2940141855172036]
Lossless image compression is required in various applications to reduce storage or transmission costs of images.
We argue that compressibility of a color image is essentially derived from the patterns in its spatial structure.
The proposed scheme first learns dictionaries of $16times16$, $8times8$, $4times4$, and $2times 2$ square pixel patterns from various datasets of binary images.
arXiv Detail & Related papers (2024-06-05T09:24:10Z) - Efficient Map Sparsification Based on 2D and 3D Discretized Grids [47.22997560184043]
As a map grows larger, more memory is required and localization becomes inefficient.
Previous map sparsification methods add a quadratic term in mixed-integer programming to enforce a uniform distribution of selected landmarks.
In this paper, we formulate map sparsification in an efficient linear form and select uniformly distributed landmarks based on 2D discretized grids.
arXiv Detail & Related papers (2023-03-20T05:49:14Z) - NeuMap: Neural Coordinate Mapping by Auto-Transdecoder for Camera
Localization [60.73541222862195]
NeuMap is an end-to-end neural mapping method for camera localization.
It encodes a whole scene into a grid of latent codes, with which a Transformer-based auto-decoder regresses 3D coordinates of query pixels.
arXiv Detail & Related papers (2022-11-21T04:46:22Z) - Neural Network Assisted Depth Map Packing for Compression Using Standard
Hardware Video Codecs [0.38073142980732994]
We propose a variable precision packing scheme assisted by a neural network model that predicts the optimal precision for each depth map given a constraint.
We demonstrate that the model yields near optimal predictions and that it can be integrated into a game engine with very low overhead.
arXiv Detail & Related papers (2022-06-30T10:46:05Z) - End-to-end lossless compression of high precision depth maps guided by
pseudo-residual [6.213322670014608]
It is urgent to explore a new compression method with better compression ratio for high precision depth maps.
We propose an end-to-end learning-based lossless compression method for high precision depth maps.
arXiv Detail & Related papers (2022-01-10T07:19:02Z) - Variable-Rate Deep Image Compression through Spatially-Adaptive Feature
Transform [58.60004238261117]
We propose a versatile deep image compression network based on Spatial Feature Transform (SFT arXiv:1804.02815)
Our model covers a wide range of compression rates using a single model, which is controlled by arbitrary pixel-wise quality maps.
The proposed framework allows us to perform task-aware image compressions for various tasks.
arXiv Detail & Related papers (2021-08-21T17:30:06Z) - Neural Network Compression Via Sparse Optimization [23.184290795230897]
We propose a model compression framework based on the recent progress on sparse optimization.
We achieve up to 7.2 and 2.9 times FLOPs reduction with the same level of evaluation of accuracy on VGG16 for CIFAR10 and ResNet50 for ImageNet.
arXiv Detail & Related papers (2020-11-10T03:03:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.