CGCV:Context Guided Correlation Volume for Optical Flow Neural Networks
- URL: http://arxiv.org/abs/2212.10174v1
- Date: Tue, 20 Dec 2022 11:24:35 GMT
- Title: CGCV:Context Guided Correlation Volume for Optical Flow Neural Networks
- Authors: Jiangpeng Li, Yan Niu
- Abstract summary: Correlation volume is the central component of optical flow computational neural models.
We propose a new Context Guided Correlation Volume (CGCV) via gating and lifting schemes.
CGCV can be universally integrated with RAFT-based flow computation methods for enhanced performance.
- Score: 1.9226937205270165
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Optical flow, which computes the apparent motion from a pair of video frames,
is a critical tool for scene motion estimation. Correlation volume is the
central component of optical flow computational neural models. It estimates the
pairwise matching costs between cross-frame features, and is then used to
decode optical flow. However, traditional correlation volume is frequently
noisy, outlier-prone, and sensitive to motion blur. We observe that, although
the recent RAFT algorithm also adopts the traditional correlation volume, its
additional context encoder provides semantically representative features to the
flow decoder, implicitly compensating for the deficiency of the correlation
volume. However, the benefits of this context encoder has been barely discussed
or exploited. In this paper, we first investigate the functionality of RAFT's
context encoder, then propose a new Context Guided Correlation Volume (CGCV)
via gating and lifting schemes. CGCV can be universally integrated with
RAFT-based flow computation methods for enhanced performance, especially
effective in the presence of motion blur, de-focus blur and atmospheric
effects. By incorporating the proposed CGCV with previous Global Motion
Aggregation (GMA) method, at a minor cost of 0.5% extra parameters, the rank of
GMA is lifted by 23 places on KITTI 2015 Leader Board, and 3 places on Sintel
Leader Board. Moreover, at a similar model size, our correlation volume
achieves competitive or superior performance to state of the art peer
supervised models that employ Transformers or Graph Reasoning, as verified by
extensive experiments.
Related papers
- OCAI: Improving Optical Flow Estimation by Occlusion and Consistency Aware Interpolation [55.676358801492114]
We propose OCAI, a method that supports robust frame ambiguities by generating intermediate video frames alongside optical flows in between.
Our evaluations demonstrate superior quality and enhanced optical flow accuracy on established benchmarks such as Sintel and KITTI.
arXiv Detail & Related papers (2024-03-26T20:23:48Z) - Motion-Aware Video Frame Interpolation [49.49668436390514]
We introduce a Motion-Aware Video Frame Interpolation (MA-VFI) network, which directly estimates intermediate optical flow from consecutive frames.
It not only extracts global semantic relationships and spatial details from input frames with different receptive fields, but also effectively reduces the required computational cost and complexity.
arXiv Detail & Related papers (2024-02-05T11:00:14Z) - GAFlow: Incorporating Gaussian Attention into Optical Flow [62.646389181507764]
We push Gaussian Attention (GA) into the optical flow models to accentuate local properties during representation learning.
We introduce a novel Gaussian-Constrained Layer (GCL) which can be easily plugged into existing Transformer blocks.
For reliable motion analysis, we provide a new Gaussian-Guided Attention Module (GGAM)
arXiv Detail & Related papers (2023-09-28T07:46:01Z) - Dynamic Frame Interpolation in Wavelet Domain [57.25341639095404]
Video frame is an important low-level computation vision task, which can increase frame rate for more fluent visual experience.
Existing methods have achieved great success by employing advanced motion models and synthesis networks.
WaveletVFI can reduce computation up to 40% while maintaining similar accuracy, making it perform more efficiently against other state-of-the-arts.
arXiv Detail & Related papers (2023-09-07T06:41:15Z) - CRAFT: Cross-Attentional Flow Transformer for Robust Optical Flow [23.457898451057275]
Optical flow estimation aims to find the 2D motion field by identifying corresponding pixels between two images.
Despite the tremendous progress of deep learning-based optical flow methods, it remains a challenge to accurately estimate large displacements with motion blur.
This is mainly because the correlation volume, the basis of pixel matching, is computed as the dot product of the convolutional features of the two images.
We propose a new architecture "CRoss-Attentional Flow Transformer" (CRAFT) to revitalize the correlation volume computation.
arXiv Detail & Related papers (2022-03-31T09:05:00Z) - Deep Calibration of Interest Rates Model [0.0]
Despite the growing use of Deep Learning, classic rate models such as CIR and the Gaussian family are still widely used.
In this paper, we propose to calibrate the five parameters of the G2++ model using Neural Networks.
arXiv Detail & Related papers (2021-10-28T14:08:45Z) - Correlate-and-Excite: Real-Time Stereo Matching via Guided Cost Volume
Excitation [65.83008812026635]
We construct Guided Cost volume Excitation (GCE) and show that simple channel excitation of cost volume guided by image can improve performance considerably.
We present an end-to-end network that we call Correlate-and-Excite (CoEx)
arXiv Detail & Related papers (2021-08-12T14:32:26Z) - Learning Optical Flow from a Few Matches [67.83633948984954]
We show that the dense correlation volume representation is redundant and accurate flow estimation can be achieved with only a fraction of elements in it.
Experiments show that our method can reduce computational cost and memory use significantly, while maintaining high accuracy.
arXiv Detail & Related papers (2021-04-05T21:44:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.