A Look at Improving Robustness in Visual-inertial SLAM by Moment
Matching
- URL: http://arxiv.org/abs/2205.13821v1
- Date: Fri, 27 May 2022 08:22:03 GMT
- Title: A Look at Improving Robustness in Visual-inertial SLAM by Moment
Matching
- Authors: Arno Solin, Rui Li, Andrea Pilzer
- Abstract summary: This paper takes a critical look at the practical implications and limitations posed by the extended Kalman filter (EKF)
We employ a moment matching (unscented Kalman filtering) approach to both visual-inertial odometry and visual SLAM.
- Score: 17.995121900076615
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The fusion of camera sensor and inertial data is a leading method for
ego-motion tracking in autonomous and smart devices. State estimation
techniques that rely on non-linear filtering are a strong paradigm for solving
the associated information fusion task. The de facto inference method in this
space is the celebrated extended Kalman filter (EKF), which relies on
first-order linearizations of both the dynamical and measurement model. This
paper takes a critical look at the practical implications and limitations posed
by the EKF, especially under faulty visual feature associations and the
presence of strong confounding noise. As an alternative, we revisit the assumed
density formulation of Bayesian filtering and employ a moment matching
(unscented Kalman filtering) approach to both visual-inertial odometry and
visual SLAM. Our results highlight important aspects in robustness both in
dynamics propagation and visual measurement updates, and we show
state-of-the-art results on EuRoC MAV drone data benchmark.
Related papers
- Localized Gaussians as Self-Attention Weights for Point Clouds Correspondence [92.07601770031236]
We investigate semantically meaningful patterns in the attention heads of an encoder-only Transformer architecture.
We find that fixing the attention weights not only accelerates the training process but also enhances the stability of the optimization.
arXiv Detail & Related papers (2024-09-20T07:41:47Z) - KFD-NeRF: Rethinking Dynamic NeRF with Kalman Filter [49.85369344101118]
We introduce KFD-NeRF, a novel dynamic neural radiance field integrated with an efficient and high-quality motion reconstruction framework based on Kalman filtering.
Our key idea is to model the dynamic radiance field as a dynamic system whose temporally varying states are estimated based on two sources of knowledge: observations and predictions.
Our KFD-NeRF demonstrates similar or even superior performance within comparable computational time and state-of-the-art view synthesis performance with thorough training.
arXiv Detail & Related papers (2024-07-18T05:48:24Z) - A Gentle Approach to Multi-Sensor Fusion Data Using Linear Kalman Filter [6.17569685604975]
This research paper delves into the Linear Kalman Filter (LKF), highlighting its importance in merging data from multiple sensors.
Our focus is on linear dynamic systems due to the LKF's assumptions about system dynamics, measurement noise, and initial conditions.
This fusion is essential for integrating diverse sensory inputs, thereby improving the accuracy and reliability of state estimations.
arXiv Detail & Related papers (2024-05-27T17:46:58Z) - Diffusion-Based Particle-DETR for BEV Perception [94.88305708174796]
Bird-Eye-View (BEV) is one of the most widely-used scene representations for visual perception in Autonomous Vehicles (AVs)
Recent diffusion-based methods offer a promising approach to uncertainty modeling for visual perception but fail to effectively detect small objects in the large coverage of the BEV.
Here, we address this problem by combining the diffusion paradigm with current state-of-the-art 3D object detectors in BEV.
arXiv Detail & Related papers (2023-12-18T09:52:14Z) - Graph Signal Diffusion Model for Collaborative Filtering [22.727100820178414]
Collaborative filtering is a critical technique in recommender systems.
Existing studies on diffusion model lack effective solutions for modeling implicit feedback.
We propose a novel Graph Signal Diffusion Model for Collaborative Filtering (named GiffCF)
arXiv Detail & Related papers (2023-11-15T07:25:14Z) - Towards Scale-Aware, Robust, and Generalizable Unsupervised Monocular
Depth Estimation by Integrating IMU Motion Dynamics [74.1720528573331]
Unsupervised monocular depth and ego-motion estimation has drawn extensive research attention in recent years.
We propose DynaDepth, a novel scale-aware framework that integrates information from vision and IMU motion dynamics.
We validate the effectiveness of DynaDepth by conducting extensive experiments and simulations on the KITTI and Make3D datasets.
arXiv Detail & Related papers (2022-07-11T07:50:22Z) - DepthFormer: Exploiting Long-Range Correlation and Local Information for
Accurate Monocular Depth Estimation [50.08080424613603]
Long-range correlation is essential for accurate monocular depth estimation.
We propose to leverage the Transformer to model this global context with an effective attention mechanism.
Our proposed model, termed DepthFormer, surpasses state-of-the-art monocular depth estimation methods with prominent margins.
arXiv Detail & Related papers (2022-03-27T05:03:56Z) - Depth-conditioned Dynamic Message Propagation for Monocular 3D Object
Detection [86.25022248968908]
We learn context- and depth-aware feature representation to solve the problem of monocular 3D object detection.
We show state-of-the-art results among the monocular-based approaches on the KITTI benchmark dataset.
arXiv Detail & Related papers (2021-03-30T16:20:24Z) - Robust Correlation Tracking via Multi-channel Fused Features and
Reliable Response Map [10.079856376445598]
This paper proposes a robust correlation tracking algorithm (RCT) based on two ideas.
First, we propose a method to fuse features in order to more naturally describe the gradient and color information of the tracked object.
Second, we present a novel strategy to significantly reduce noise in the response map and therefore ease the problem of model drift.
arXiv Detail & Related papers (2020-11-25T07:15:03Z) - Robust Ego and Object 6-DoF Motion Estimation and Tracking [5.162070820801102]
This paper proposes a robust solution to achieve accurate estimation and consistent track-ability for dynamic multi-body visual odometry.
A compact and effective framework is proposed leveraging recent advances in semantic instance-level segmentation and accurate optical flow estimation.
A novel formulation, jointly optimizing SE(3) motion and optical flow is introduced that improves the quality of the tracked points and the motion estimation accuracy.
arXiv Detail & Related papers (2020-07-28T05:12:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.