SSTM: Spatiotemporal Recurrent Transformers for Multi-frame Optical Flow
Estimation
- URL: http://arxiv.org/abs/2304.14418v1
- Date: Wed, 26 Apr 2023 23:39:40 GMT
- Title: SSTM: Spatiotemporal Recurrent Transformers for Multi-frame Optical Flow
Estimation
- Authors: Fisseha Admasu Ferede, Madhusudhanan Balasubramanian
- Abstract summary: In optical flow estimates in and near cluded regions, and out-of-boundary regions are two of the current significant limitations of optical flow estimation algorithms.
Recent state-of-the-art optical flow estimation algorithms are two-frame based methods where optical flow is estimated sequentially for each consecutive image pair in a sequence.
We propose a learning-based multi-frame optical flow estimation method that estimates two or more consecutive optical flows in parallel from multi-frame image sequences.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Inaccurate optical flow estimates in and near occluded regions, and
out-of-boundary regions are two of the current significant limitations of
optical flow estimation algorithms. Recent state-of-the-art optical flow
estimation algorithms are two-frame based methods where optical flow is
estimated sequentially for each consecutive image pair in a sequence. While
this approach gives good flow estimates, it fails to generalize optical flows
in occluded regions mainly due to limited local evidence regarding moving
elements in a scene. In this work, we propose a learning-based multi-frame
optical flow estimation method that estimates two or more consecutive optical
flows in parallel from multi-frame image sequences. Our underlying hypothesis
is that by understanding temporal scene dynamics from longer sequences with
more than two frames, we can characterize pixel-wise dependencies in a larger
spatiotemporal domain, generalize complex motion patterns and thereby improve
the accuracy of optical flow estimates in occluded regions. We present
learning-based spatiotemporal recurrent transformers for multi-frame based
optical flow estimation (SSTMs). Our method utilizes 3D Convolutional Gated
Recurrent Units (3D-ConvGRUs) and spatiotemporal transformers to learn
recurrent space-time motion dynamics and global dependencies in the scene and
provide a generalized optical flow estimation. When compared with recent
state-of-the-art two-frame and multi-frame methods on real world and synthetic
datasets, performance of the SSTMs were significantly higher in occluded and
out-of-boundary regions. Among all published state-of-the-art multi-frame
methods, SSTM achieved state-of the-art results on the Sintel Final and
KITTI2015 benchmark datasets.
Related papers
- Robust Optical Flow Computation: A Higher-Order Differential Approach [0.0]
This research proposes an innovative algorithm for optical flow computation, utilizing the higher precision of second-order Taylor series approximation.
An impressive showcase of the algorithm's capabilities emerges through its performance on optical flow benchmarks such as KITTI and Middlebury.
arXiv Detail & Related papers (2024-10-12T15:20:11Z) - OCAI: Improving Optical Flow Estimation by Occlusion and Consistency Aware Interpolation [55.676358801492114]
We propose OCAI, a method that supports robust frame ambiguities by generating intermediate video frames alongside optical flows in between.
Our evaluations demonstrate superior quality and enhanced optical flow accuracy on established benchmarks such as Sintel and KITTI.
arXiv Detail & Related papers (2024-03-26T20:23:48Z) - Motion-Aware Video Frame Interpolation [49.49668436390514]
We introduce a Motion-Aware Video Frame Interpolation (MA-VFI) network, which directly estimates intermediate optical flow from consecutive frames.
It not only extracts global semantic relationships and spatial details from input frames with different receptive fields, but also effectively reduces the required computational cost and complexity.
arXiv Detail & Related papers (2024-02-05T11:00:14Z) - STint: Self-supervised Temporal Interpolation for Geospatial Data [0.0]
Supervised and unsupervised techniques have demonstrated the potential for temporal of video data.
Most prevailing temporal techniques hinge on optical flow, which encodes the motion of pixels between video frames.
In this work, we propose an unsupervised temporal technique, which does not rely on ground truth data or require any motion information like optical flow.
arXiv Detail & Related papers (2023-08-31T18:04:50Z) - Learning Dense and Continuous Optical Flow from an Event Camera [28.77846425802558]
Event cameras such as DAVIS can simultaneously output high temporal resolution events and low frame-rate intensity images.
Most of the existing optical flow estimation methods are based on two consecutive image frames and can only estimate discrete flow at a fixed time interval.
We propose a novel deep learning-based dense and continuous optical flow estimation framework from a single image with event streams.
arXiv Detail & Related papers (2022-11-16T17:53:18Z) - Optical-Flow-Reuse-Based Bidirectional Recurrent Network for Space-Time
Video Super-Resolution [52.899234731501075]
Space-time video super-resolution (ST-VSR) simultaneously increases the spatial resolution and frame rate for a given video.
Existing methods typically suffer from difficulties in how to efficiently leverage information from a large range of neighboring frames.
We propose a coarse-to-fine bidirectional recurrent neural network instead of using ConvLSTM to leverage knowledge between adjacent frames.
arXiv Detail & Related papers (2021-10-13T15:21:30Z) - Dense Optical Flow from Event Cameras [55.79329250951028]
We propose to incorporate feature correlation and sequential processing into dense optical flow estimation from event cameras.
Our proposed approach computes dense optical flow and reduces the end-point error by 23% on MVSEC.
arXiv Detail & Related papers (2021-08-24T07:39:08Z) - TimeLens: Event-based Video Frame Interpolation [54.28139783383213]
We introduce Time Lens, a novel indicates equal contribution method that leverages the advantages of both synthesis-based and flow-based approaches.
We show an up to 5.21 dB improvement in terms of PSNR over state-of-the-art frame-based and event-based methods.
arXiv Detail & Related papers (2021-06-14T10:33:47Z) - PCA Event-Based Otical Flow for Visual Odometry [0.0]
We present a Principal Component Analysis approach to the problem of event-based optical flow estimation.
We show that the best variant of our proposed method, dedicated to the real-time context of visual odometry, is about two times faster compared to state-of-the-art implementations.
arXiv Detail & Related papers (2021-05-08T18:30:44Z) - STaRFlow: A SpatioTemporal Recurrent Cell for Lightweight Multi-Frame
Optical Flow Estimation [64.99259320624148]
We present a new lightweight CNN-based algorithm for multi-frame optical flow estimation.
The resulting STaRFlow algorithm gives state-of-the-art performances on MPI Sintel and Kitti2015.
arXiv Detail & Related papers (2020-07-10T17:01:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.