Combining 3D Model Contour Energy and Keypoints for Object Tracking
- URL: http://arxiv.org/abs/2002.01379v1
- Date: Tue, 4 Feb 2020 15:53:26 GMT
- Title: Combining 3D Model Contour Energy and Keypoints for Object Tracking
- Authors: Bogdan Bugaev, Anton Kryshchenko, Roman Belov
- Abstract summary: We present a new combined approach for monocular model-based 3D tracking.
A preliminary object pose is estimated by using a keypoint-based technique.
The pose is then refined by optimizing the contour energy function.
- Score: 2.5782420501870287
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a new combined approach for monocular model-based 3D tracking. A
preliminary object pose is estimated by using a keypoint-based technique. The
pose is then refined by optimizing the contour energy function. The energy
determines the degree of correspondence between the contour of the model
projection and the image edges. It is calculated based on both the intensity
and orientation of the raw image gradient. For optimization, we propose a
technique and search area constraints that allow overcoming the local optima
and taking into account information obtained through keypoint-based pose
estimation. Owing to its combined nature, our method eliminates numerous issues
of keypoint-based and edge-based approaches. We demonstrate the efficiency of
our method by comparing it with state-of-the-art methods on a public benchmark
dataset that includes videos with various lighting conditions, movement
patterns, and speed.
Related papers
- KRONC: Keypoint-based Robust Camera Optimization for 3D Car Reconstruction [58.04846444985808]
This paper introduces KRONC, a novel approach aimed at inferring view poses by leveraging prior knowledge about the object to reconstruct and its representation through semantic keypoints.
With a focus on vehicle scenes, KRONC is able to estimate the position of the views as a solution to a light optimization problem targeting the convergence of keypoints' back-projections to a singular point.
arXiv Detail & Related papers (2024-09-09T08:08:05Z) - FlowMap: High-Quality Camera Poses, Intrinsics, and Depth via Gradient Descent [19.977807508281835]
FlowMap is an end-to-end differentiable method that solves for precise camera poses, camera intrinsics, and per-frame dense depth of a video sequence.
Our method performs per-video gradient-descent minimization of a simple least-squares objective.
We empirically show that camera parameters and dense depth recovered by our method enable photo-realistic novel view synthesis on 360-degree trajectories.
arXiv Detail & Related papers (2024-04-23T17:46:50Z) - Towards Scalable Multi-View Reconstruction of Geometry and Materials [27.660389147094715]
We propose a novel method for joint recovery of camera pose, object geometry and spatially-varying Bidirectional Reflectance Distribution Function (svBRDF) of 3D scenes.
The input are high-resolution RGBD images captured by a mobile, hand-held capture system with point lights for active illumination.
arXiv Detail & Related papers (2023-06-06T15:07:39Z) - 6DOF Pose Estimation of a 3D Rigid Object based on Edge-enhanced Point
Pair Features [20.33119373900788]
We propose an efficient 6D pose estimation method based on the point pair feature (PPF) framework.
A pose hypothesis validation approach is proposed to resolve the symmetric ambiguity by calculating edge matching degree.
arXiv Detail & Related papers (2022-09-17T07:05:50Z) - RelPose: Predicting Probabilistic Relative Rotation for Single Objects
in the Wild [73.1276968007689]
We describe a data-driven method for inferring the camera viewpoints given multiple images of an arbitrary object.
We show that our approach outperforms state-of-the-art SfM and SLAM methods given sparse images on both seen and unseen categories.
arXiv Detail & Related papers (2022-08-11T17:59:59Z) - Leveraging Monocular Disparity Estimation for Single-View Reconstruction [8.583436410810203]
We leverage advances in monocular depth estimation to obtain disparity maps.
We transform 2D normalized disparity maps into 3D point clouds by solving an optimization on the relevant camera parameters.
arXiv Detail & Related papers (2022-07-01T03:05:40Z) - Semantic keypoint-based pose estimation from single RGB frames [64.80395521735463]
We present an approach to estimating the continuous 6-DoF pose of an object from a single RGB image.
The approach combines semantic keypoints predicted by a convolutional network (convnet) with a deformable shape model.
We show that our approach can accurately recover the 6-DoF object pose for both instance- and class-based scenarios.
arXiv Detail & Related papers (2022-04-12T15:03:51Z) - A Model for Multi-View Residual Covariances based on Perspective
Deformation [88.21738020902411]
We derive a model for the covariance of the visual residuals in multi-view SfM, odometry and SLAM setups.
We validate our model with synthetic and real data and integrate it into photometric and feature-based Bundle Adjustment.
arXiv Detail & Related papers (2022-02-01T21:21:56Z) - Learning Stereopsis from Geometric Synthesis for 6D Object Pose
Estimation [11.999630902627864]
Current monocular-based 6D object pose estimation methods generally achieve less competitive results than RGBD-based methods.
This paper proposes a 3D geometric volume based pose estimation method with a short baseline two-view setting.
Experiments show that our method outperforms state-of-the-art monocular-based methods, and is robust in different objects and scenes.
arXiv Detail & Related papers (2021-09-25T02:55:05Z) - Learning Monocular Depth in Dynamic Scenes via Instance-Aware Projection
Consistency [114.02182755620784]
We present an end-to-end joint training framework that explicitly models 6-DoF motion of multiple dynamic objects, ego-motion and depth in a monocular camera setup without supervision.
Our framework is shown to outperform the state-of-the-art depth and motion estimation methods.
arXiv Detail & Related papers (2021-02-04T14:26:42Z) - Nothing But Geometric Constraints: A Model-Free Method for Articulated
Object Pose Estimation [89.82169646672872]
We propose an unsupervised vision-based system to estimate the joint configurations of the robot arm from a sequence of RGB or RGB-D images without knowing the model a priori.
We combine a classical geometric formulation with deep learning and extend the use of epipolar multi-rigid-body constraints to solve this task.
arXiv Detail & Related papers (2020-11-30T20:46:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.