Learnable Cost Volume Using the Cayley Representation
- URL: http://arxiv.org/abs/2007.11431v1
- Date: Tue, 21 Jul 2020 01:59:36 GMT
- Title: Learnable Cost Volume Using the Cayley Representation
- Authors: Taihong Xiao, Jinwei Yuan, Deqing Sun, Qifei Wang, Xin-Yu Zhang, Kehan
Xu, Ming-Hsuan Yang
- Abstract summary: We propose a learnable cost volume (LCV) for optical flow estimation models.
The proposed LCV is a lightweight module and can be easily plugged into existing models to replace the vanilla cost volume.
Experimental results show that the LCV module not only improves the accuracy of state-of-the-art models on standard benchmarks, but also promotes their robustness against illumination change, noises, and adversarial perturbations of the input signals.
- Score: 67.19770048548232
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Cost volume is an essential component of recent deep models for optical flow
estimation and is usually constructed by calculating the inner product between
two feature vectors. However, the standard inner product in the commonly-used
cost volume may limit the representation capacity of flow models because it
neglects the correlation among different channel dimensions and weighs each
dimension equally. To address this issue, we propose a learnable cost volume
(LCV) using an elliptical inner product, which generalizes the standard inner
product by a positive definite kernel matrix. To guarantee its positive
definiteness, we perform spectral decomposition on the kernel matrix and
re-parameterize it via the Cayley representation. The proposed LCV is a
lightweight module and can be easily plugged into existing models to replace
the vanilla cost volume. Experimental results show that the LCV module not only
improves the accuracy of state-of-the-art models on standard benchmarks, but
also promotes their robustness against illumination change, noises, and
adversarial perturbations of the input signals.
Related papers
- Efficient Variational Quantum Linear Solver for Structured Sparse Matrices [0.6138671548064355]
We show that by using an alternate basis one can better exploit the sparsity and underlying structure of matrix.
We employ the concept of unitary completion to design efficient quantum circuits for computing the global/local VQLS cost functions.
arXiv Detail & Related papers (2024-04-25T19:22:05Z) - Adaptive Computation Modules: Granular Conditional Computation For
Efficient Inference [13.000030080938078]
computational cost of transformer models makes them inefficient in low-latency or low-power applications.
We introduce the Adaptive Computation Module (ACM), a generic module that dynamically adapts its computational load to match the estimated difficulty of the input on a per-token basis.
Our evaluation of transformer models in computer vision and speech recognition demonstrates that substituting layers with ACMs significantly reduces inference costs without degrading the downstream accuracy for a wide interval of user-defined budgets.
arXiv Detail & Related papers (2023-12-15T20:39:43Z) - LQ-LoRA: Low-rank Plus Quantized Matrix Decomposition for Efficient Language Model Finetuning [66.85589263870702]
Our approach uses an iterative algorithm to decompose each pretrained matrix into a high-precision low-rank component and a memory-efficient quantized component.
Experiments on finetuning RoBERTa and LLaMA-2 demonstrate that our low-rank plus quantized matrix decomposition approach (LQ-LoRA) outperforms strong QLoRA and GPTQ-LoRA baselines.
arXiv Detail & Related papers (2023-11-20T18:57:41Z) - Disentanglement via Latent Quantization [60.37109712033694]
In this work, we construct an inductive bias towards encoding to and decoding from an organized latent space.
We demonstrate the broad applicability of this approach by adding it to both basic data-re (vanilla autoencoder) and latent-reconstructing (InfoGAN) generative models.
arXiv Detail & Related papers (2023-05-28T06:30:29Z) - Sketching as a Tool for Understanding and Accelerating Self-attention
for Long Sequences [52.6022911513076]
Transformer-based models are not efficient in processing long sequences due to the quadratic space and time complexity of the self-attention modules.
We propose Linformer and Informer to reduce the quadratic complexity to linear (modulo logarithmic factors) via low-dimensional projection and row selection.
Based on the theoretical analysis, we propose Skeinformer to accelerate self-attention and further improve the accuracy of matrix approximation to self-attention.
arXiv Detail & Related papers (2021-12-10T06:58:05Z) - Robust and Provably Monotonic Networks [0.0]
We present a new method to constrain the Lipschitz constant of dense deep learning models.
We show how the algorithm was used to train a powerful, robust, and interpretable discriminator for heavy-flavor decays in the LHCb realtime data-processing system.
arXiv Detail & Related papers (2021-11-30T19:01:32Z) - Learning Optical Flow from a Few Matches [67.83633948984954]
We show that the dense correlation volume representation is redundant and accurate flow estimation can be achieved with only a fraction of elements in it.
Experiments show that our method can reduce computational cost and memory use significantly, while maintaining high accuracy.
arXiv Detail & Related papers (2021-04-05T21:44:00Z) - Understanding Implicit Regularization in Over-Parameterized Single Index
Model [55.41685740015095]
We design regularization-free algorithms for the high-dimensional single index model.
We provide theoretical guarantees for the induced implicit regularization phenomenon.
arXiv Detail & Related papers (2020-07-16T13:27:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.