SVQ: Sparse Vector Quantization for Spatiotemporal Forecasting
- URL: http://arxiv.org/abs/2312.03406v3
- Date: Thu, 8 Feb 2024 02:55:08 GMT
- Title: SVQ: Sparse Vector Quantization for Spatiotemporal Forecasting
- Authors: Chao Chen, Tian Zhou, Yanjun Zhao, Hui Liu, Liang Sun, Rong Jin
- Abstract summary: We introduce Sparse Regression-based Vector Quantization (SVQ), a novel technique that leverages sparse regression for succinct representation.
In video prediction-Human, KTH, and KittiCaltech-it reduces MAE by an average of 9.4% and improves image quality by 17.3%.
Our empirical studies on five benchmark datasets demonstrate that SVQ state-of-the-art results.
- Score: 23.38628640665113
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Spatio-temporal forecasting, pivotal in numerous fields, hinges on the
delicate equilibrium between isolating nuanced patterns and sifting out noise.
To tackle this, we introduce Sparse Regression-based Vector Quantization (SVQ),
a novel technique that leverages sparse regression for succinct representation,
an approach theoretically and practically favored over classical
clustering-based vector quantization methods. This approach preserves critical
details from the original vectors using a regression model while filtering out
noise via sparse design. Moreover, we approximate the sparse regression process
using a blend of a two-layer MLP and an extensive codebook. This approach not
only substantially cuts down on computational costs but also grants SVQ
differentiability and training simplicity, resulting in a notable enhancement
of performance. Our empirical studies on five spatial-temporal benchmark
datasets demonstrate that SVQ achieves state-of-the-art results. Specifically,
on the WeatherBench-S temperature dataset, SVQ improves the top baseline by
7.9%. In video prediction benchmarks-Human, KTH, and KittiCaltech-it reduces
MAE by an average of 9.4% and improves image quality by 17.3% (LPIPS).
Related papers
- AffineQuant: Affine Transformation Quantization for Large Language Models [58.45460102764]
Post-Training Quantization (PTQ) has emerged as a subject of considerable interest due to its compression efficiency and cost-effectiveness in the context of training.
Existing PTQ methods for Large-scale Language Models (LLMs) limit the optimization scope to scaling transformations between pre- and post-quantization weights.
In this paper, we advocate for the direct optimization using equivalent Affine transformations in PTQ (AffineQuant)
arXiv Detail & Related papers (2024-03-19T08:40:21Z) - Towards Accurate Post-training Quantization for Reparameterized Models [6.158896686945439]
Current Post-training Quantization (PTQ) methods often lead to significant accuracy degradation.
This is primarily caused by channel-specific and sample-specific outliers.
We propose RepAPQ, a novel framework that preserves the accuracy of quantized reparameterization models.
arXiv Detail & Related papers (2024-02-25T15:42:12Z) - Stochastic Gradient Descent for Gaussian Processes Done Right [86.83678041846971]
We show that when emphdone right -- by which we mean using specific insights from optimisation and kernel communities -- gradient descent is highly effective.
We introduce a emphstochastic dual descent algorithm, explain its design in an intuitive manner and illustrate the design choices.
Our method places Gaussian process regression on par with state-of-the-art graph neural networks for molecular binding affinity prediction.
arXiv Detail & Related papers (2023-10-31T16:15:13Z) - Balancing Computational Efficiency and Forecast Error in Machine
Learning-based Time-Series Forecasting: Insights from Live Experiments on
Meteorological Nowcasting [0.0]
This paper addresses the relationship between computational cost and forecast error using meteorological nowcasting.
During a 5-day live experiment, 4000 data sources were streamed for training and inferencing 144 models per hour.
Results indicate that using the Variance Horizon reduced computational usage by more than 50%, while increasing between 0-15% in error.
arXiv Detail & Related papers (2023-09-26T19:10:00Z) - QuantEase: Optimization-based Quantization for Language Models [17.333778751252392]
This work introduces Quantization (PTQ) of various quantization layers from recent advances of Large Language Models (LLMs)
Our CD-based approach features straightforward updates, relying solely on vector operations.
We also explore an outlier approach, allowing for retaining significant weights (outoutliers) with complete precision.
arXiv Detail & Related papers (2023-09-05T01:39:09Z) - S-HR-VQVAE: Sequential Hierarchical Residual Learning Vector Quantized Variational Autoencoder for Video Prediction [16.14728977379756]
We put forth a novel model that combines a novel residual vector learning quantized variational autoencoder (HR-VQE) and a hierarchical autoregressive vector predictive model (AST-PM)
We show that our model compares favorably against state-of-the-art video prediction techniques both in quantitative and qualitative evaluations despite a much smaller model size.
arXiv Detail & Related papers (2023-07-13T11:58:27Z) - Solving Oscillation Problem in Post-Training Quantization Through a
Theoretical Perspective [74.48124653728422]
Post-training quantization (PTQ) is widely regarded as one of the most efficient compression methods practically.
We argue that an overlooked problem of oscillation is in the PTQ methods.
arXiv Detail & Related papers (2023-03-21T14:52:52Z) - Training Discrete Deep Generative Models via Gapped Straight-Through
Estimator [72.71398034617607]
We propose a Gapped Straight-Through ( GST) estimator to reduce the variance without incurring resampling overhead.
This estimator is inspired by the essential properties of Straight-Through Gumbel-Softmax.
Experiments demonstrate that the proposed GST estimator enjoys better performance compared to strong baselines on two discrete deep generative modeling tasks.
arXiv Detail & Related papers (2022-06-15T01:46:05Z) - Why Approximate Matrix Square Root Outperforms Accurate SVD in Global
Covariance Pooling? [59.820507600960745]
We propose a new GCP meta-layer that uses SVD in the forward pass, and Pad'e Approximants in the backward propagation to compute the gradients.
The proposed meta-layer has been integrated into different CNN models and achieves state-of-the-art performances on both large-scale and fine-grained datasets.
arXiv Detail & Related papers (2021-05-06T08:03:45Z) - An Efficient Statistical-based Gradient Compression Technique for
Distributed Training Systems [77.88178159830905]
Sparsity-Inducing Distribution-based Compression (SIDCo) is a threshold-based sparsification scheme that enjoys similar threshold estimation quality to deep gradient compression (DGC)
Our evaluation shows SIDCo speeds up training by up to 41:7%, 7:6%, and 1:9% compared to the no-compression baseline, Topk, and DGC compressors, respectively.
arXiv Detail & Related papers (2021-01-26T13:06:00Z) - Quantile Surfaces -- Generalizing Quantile Regression to Multivariate
Targets [4.979758772307178]
Our approach is based on an extension of single-output quantile regression (QR) to multivariate-targets, called quantile surfaces (QS)
We present a novel two-stage process: In the first stage, we perform a deterministic point forecast (i.e., central tendency estimation)
Subsequently, we model the prediction uncertainty using QS involving neural networks called quantile surface regression neural networks (QSNN)
We evaluate our novel approach on synthetic data and two currently researched real-world challenges in two different domains: First, probabilistic forecasting for renewable energy power generation, second, short-term cyclists trajectory forecasting for
arXiv Detail & Related papers (2020-09-29T16:35:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.