FLIGHT: Fibonacci Lattice-based Inference for Geometric Heading in real-Time
- URL: http://arxiv.org/abs/2602.23115v1
- Date: Thu, 26 Feb 2026 15:27:49 GMT
- Title: FLIGHT: Fibonacci Lattice-based Inference for Geometric Heading in real-Time
- Authors: David Dirnfeld, Fabien Delattre, Pedro Miraldo, Erik Learned-Miller,
- Abstract summary: Estimating camera motion from monocular video is a fundamental problem in computer vision.<n>Existing methods that recover the camera's heading under known rotation tend to perform well in low-noise, low-outlier conditions.<n>We propose a novel generalization of the Hough transform on the unit sphere to estimate the camera's heading.
- Score: 7.517221623631364
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Estimating camera motion from monocular video is a fundamental problem in computer vision, central to tasks such as SLAM, visual odometry, and structure-from-motion. Existing methods that recover the camera's heading under known rotation, whether from an IMU or an optimization algorithm, tend to perform well in low-noise, low-outlier conditions, but often decrease in accuracy or become computationally expensive as noise and outlier levels increase. To address these limitations, we propose a novel generalization of the Hough transform on the unit sphere (S(2)) to estimate the camera's heading. First, the method extracts correspondences between two frames and generates a great circle of directions compatible with each pair of correspondences. Then, by discretizing the unit sphere using a Fibonacci lattice as bin centers, each great circle casts votes for a range of directions, ensuring that features unaffected by noise or dynamic objects vote consistently for the correct motion direction. Experimental results on three datasets demonstrate that the proposed method is on the Pareto frontier of accuracy versus efficiency. Additionally, experiments on SLAM show that the proposed method reduces RMSE by correcting the heading during camera pose initialization.
Related papers
- Dynamic View Synthesis from Small Camera Motion Videos [56.359460602781304]
We present a novel view synthesis for dynamic $3$D scenes based on distribution-based depth regularization.<n>We also introduce constraints that enforce the volume density of spatial points before the object boundary along the ray to be near zero, ensuring that our model learns the correct geometry of the scene.<n>We conduct extensive experiments to demonstrate the effectiveness of our approach in representing scenes with small camera motion input, and our results compare favorably to state-of-the-art methods.
arXiv Detail & Related papers (2025-06-29T09:17:55Z) - ESVO2: Direct Visual-Inertial Odometry with Stereo Event Cameras [41.992980062962495]
Event-based visual odometry aims at solving tracking and mapping subproblems (typically in parallel)<n>We build an event-based stereo visual-inertial odometry system on top of a direct pipeline.<n>The resulting system scales well with modern high-resolution event cameras.
arXiv Detail & Related papers (2024-10-12T05:35:27Z) - An Accurate and Real-time Relative Pose Estimation from Triple Point-line Images by Decoupling Rotation and Translation [10.05584976985694]
3D-2D constraints provided by line features have been widely used in Visual Odometry (VO) and Structure-from-Motion (SfM) systems.<n>We propose a novel three-view pose solver based on rotation-translation decoupled estimation.
arXiv Detail & Related papers (2024-03-18T10:21:05Z) - DFR: Depth from Rotation by Uncalibrated Image Rectification with
Latitudinal Motion Assumption [6.369764116066747]
We propose Depth-from-Rotation (DfR), a novel image rectification solution for uncalibrated rotating cameras.
Specifically, we model the motion of a rotating camera as the camera rotates on a sphere with fixed latitude.
We derive a 2-point analytical solver from directly computing the rectified transformations on the two images.
arXiv Detail & Related papers (2023-07-11T09:11:22Z) - Detecting Rotated Objects as Gaussian Distributions and Its 3-D
Generalization [81.29406957201458]
Existing detection methods commonly use a parameterized bounding box (BBox) to model and detect (horizontal) objects.
We argue that such a mechanism has fundamental limitations in building an effective regression loss for rotation detection.
We propose to model the rotated objects as Gaussian distributions.
We extend our approach from 2-D to 3-D with a tailored algorithm design to handle the heading estimation.
arXiv Detail & Related papers (2022-09-22T07:50:48Z) - ParticleSfM: Exploiting Dense Point Trajectories for Localizing Moving
Cameras in the Wild [57.37891682117178]
We present a robust dense indirect structure-from-motion method for videos that is based on dense correspondence from pairwise optical flow.
A novel neural network architecture is proposed for processing irregular point trajectory data.
Experiments on MPI Sintel dataset show that our system produces significantly more accurate camera trajectories.
arXiv Detail & Related papers (2022-07-19T09:19:45Z) - Visual Odometry with an Event Camera Using Continuous Ray Warping and
Volumetric Contrast Maximization [31.627936023222052]
We present a new solution to tracking and mapping with an event camera.
The motion of the camera contains both rotation and translation, and the displacements happen in an arbitrarily structured environment.
We introduce a new solution to this problem by performing contrast in 3D.
The practical validity of our approach is supported by an application to AGV motion estimation and 3D reconstruction with a single vehicle-mounted event camera.
arXiv Detail & Related papers (2021-07-07T04:32:57Z) - Leveraging Spatial and Photometric Context for Calibrated Non-Lambertian
Photometric Stereo [61.6260594326246]
We introduce an efficient fully-convolutional architecture that can leverage both spatial and photometric context simultaneously.
Using separable 4D convolutions and 2D heat-maps reduces the size and makes more efficient.
arXiv Detail & Related papers (2021-03-22T18:06:58Z) - Calibrated and Partially Calibrated Semi-Generalized Homographies [65.29477277713205]
We propose the first minimal solutions for estimating the semi-generalized homography given a perspective and a generalized camera.
The proposed solvers are stable and efficient as demonstrated by a number of synthetic and real-world experiments.
arXiv Detail & Related papers (2021-03-11T08:56:24Z) - Canny-VO: Visual Odometry with RGB-D Cameras based on Geometric 3D-2D
Edge Alignment [85.32080531133799]
This paper reviews the classical problem of free-form curve registration and applies it to an efficient RGBD visual odometry system called Canny-VO.
Two replacements for the distance transformation commonly used in edge registration are proposed: Approximate Nearest Neighbour Fields and Oriented Nearest Neighbour Fields.
3D2D edge alignment benefits from these alternative formulations in terms of both efficiency and accuracy.
arXiv Detail & Related papers (2020-12-15T11:42:17Z) - Monocular Rotational Odometry with Incremental Rotation Averaging and
Loop Closure [35.467052373502575]
Estimating absolute camera orientations is essential for attitude estimation tasks.
We devise a fast algorithm to accurately estimate camera orientations with 2D-2D feature matches alone.
Underpinning our system is a new incremental rotation averaging method for fast and constant time iterative updating.
arXiv Detail & Related papers (2020-10-05T09:19:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.