A Framework for 3D Tracking of Frontal Dynamic Objects in Autonomous
Cars
- URL: http://arxiv.org/abs/2103.13430v1
- Date: Wed, 24 Mar 2021 18:21:29 GMT
- Title: A Framework for 3D Tracking of Frontal Dynamic Objects in Autonomous
Cars
- Authors: Faraz Lotfi, Hamid D. Taghirad
- Abstract summary: In this paper, the YOLOv3 approach is utilized beside an OpenCV tracker to elicit features from an image.
To obtain the lateral and longitudinal distances, a nonlinear SFM model is considered alongside a state-dependent Riccati equation filter.
A switching method in the form of switching estimation error covariance is proposed to enhance the robust performance of the SDRE filter.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Both recognition and 3D tracking of frontal dynamic objects are crucial
problems in an autonomous vehicle, while depth estimation as an essential issue
becomes a challenging problem using a monocular camera. Since both camera and
objects are moving, the issue can be formed as a structure from motion (SFM)
problem. In this paper, to elicit features from an image, the YOLOv3 approach
is utilized beside an OpenCV tracker. Subsequently, to obtain the lateral and
longitudinal distances, a nonlinear SFM model is considered alongside a
state-dependent Riccati equation (SDRE) filter and a newly developed
observation model. Additionally, a switching method in the form of switching
estimation error covariance is proposed to enhance the robust performance of
the SDRE filter. The stability analysis of the presented filter is conducted on
a class of discrete nonlinear systems. Furthermore, the ultimate bound of
estimation error caused by model uncertainties is analytically obtained to
investigate the switching significance. Simulations are reported to validate
the performance of the switched SDRE filter. Finally, real-time experiments are
performed through a multi-thread framework implemented on a Jetson TX2 board,
while radar data is used for the evaluation.
Related papers
- 3D Equivariant Pose Regression via Direct Wigner-D Harmonics Prediction [50.07071392673984]
Existing methods learn 3D rotations parametrized in the spatial domain using angles or quaternions.
We propose a frequency-domain approach that directly predicts Wigner-D coefficients for 3D rotation regression.
Our method achieves state-of-the-art results on benchmarks such as ModelNet10-SO(3) and PASCAL3D+.
arXiv Detail & Related papers (2024-11-01T12:50:38Z) - Track Initialization and Re-Identification for~3D Multi-View Multi-Object Tracking [12.389483990547223]
We propose a 3D multi-object tracking (MOT) solution using only 2D detections from monocular cameras.
We exploit the 2D detections and extracted features from multiple cameras to provide a better approximation of the multi-object filtering density.
arXiv Detail & Related papers (2024-05-28T21:36:16Z) - Pedestrian Tracking with Monocular Camera using Unconstrained 3D Motion Model [0.0]
It is assumed that the extent of the moving object can be described via known statistics in 3D, such as pedestrian height.
A nonlinear filter for this model is implemented using the unscented Kalman filter (UKF) and tested using the publicly available MOT-17 dataset.
The proposed solution yields promising results in 3D while maintaining perfect results when projected into the 2D image.
arXiv Detail & Related papers (2024-03-18T17:13:18Z) - Mip-Splatting: Alias-free 3D Gaussian Splatting [52.366815964832426]
3D Gaussian Splatting has demonstrated impressive novel view synthesis results, reaching high fidelity and efficiency.
Strong artifacts can be observed when changing the sampling rate, eg, by changing focal length or camera distance.
We find that the source for this phenomenon can be attributed to the lack of 3D frequency constraints and the usage of a 2D dilation filter.
arXiv Detail & Related papers (2023-11-27T13:03:09Z) - D-SCo: Dual-Stream Conditional Diffusion for Monocular Hand-Held Object Reconstruction [74.49121940466675]
We introduce centroid-fixed dual-stream conditional diffusion for monocular hand-held object reconstruction.
First, to avoid the object centroid from deviating, we utilize a novel hand-constrained centroid fixing paradigm.
Second, we introduce a dual-stream denoiser to semantically and geometrically model hand-object interactions.
arXiv Detail & Related papers (2023-11-23T20:14:50Z) - Detecting Rotated Objects as Gaussian Distributions and Its 3-D
Generalization [81.29406957201458]
Existing detection methods commonly use a parameterized bounding box (BBox) to model and detect (horizontal) objects.
We argue that such a mechanism has fundamental limitations in building an effective regression loss for rotation detection.
We propose to model the rotated objects as Gaussian distributions.
We extend our approach from 2-D to 3-D with a tailored algorithm design to handle the heading estimation.
arXiv Detail & Related papers (2022-09-22T07:50:48Z) - CAMO-MOT: Combined Appearance-Motion Optimization for 3D Multi-Object
Tracking with Camera-LiDAR Fusion [34.42289908350286]
3D Multi-object tracking (MOT) ensures consistency during continuous dynamic detection.
It can be challenging to accurately track the irregular motion of objects for LiDAR-based methods.
We propose a novel camera-LiDAR fusion 3D MOT framework based on the Combined Appearance-Motion Optimization (CAMO-MOT)
arXiv Detail & Related papers (2022-09-06T14:41:38Z) - RNNPose: Recurrent 6-DoF Object Pose Refinement with Robust
Correspondence Field Estimation and Pose Optimization [46.144194562841435]
We propose a framework based on a recurrent neural network (RNN) for object pose refinement.
The problem is formulated as a non-linear least squares problem based on the estimated correspondence field.
The correspondence field estimation and pose refinement are conducted alternatively in each iteration to recover accurate object poses.
arXiv Detail & Related papers (2022-03-24T06:24:55Z) - PLUME: Efficient 3D Object Detection from Stereo Images [95.31278688164646]
Existing methods tackle the problem in two steps: first depth estimation is performed, a pseudo LiDAR point cloud representation is computed from the depth estimates, and then object detection is performed in 3D space.
We propose a model that unifies these two tasks in the same metric space.
Our approach achieves state-of-the-art performance on the challenging KITTI benchmark, with significantly reduced inference time compared with existing methods.
arXiv Detail & Related papers (2021-01-17T05:11:38Z) - Reinforced Axial Refinement Network for Monocular 3D Object Detection [160.34246529816085]
Monocular 3D object detection aims to extract the 3D position and properties of objects from a 2D input image.
Conventional approaches sample 3D bounding boxes from the space and infer the relationship between the target object and each of them, however, the probability of effective samples is relatively small in the 3D space.
We propose to start with an initial prediction and refine it gradually towards the ground truth, with only one 3d parameter changed in each step.
This requires designing a policy which gets a reward after several steps, and thus we adopt reinforcement learning to optimize it.
arXiv Detail & Related papers (2020-08-31T17:10:48Z) - Active Depth Estimation: Stability Analysis and its Applications [18.582561853987034]
This paper focuses on the theoretical properties of the Structure-from-Motion (SfM) scheme.
The term incremental stands for estimating the 3D structure of the scene over a chronological sequence of image frames.
By analyzing the convergence of the estimator using the Lyapunov theory, we relax the constraints on the projection of the 3D point in the image plane.
arXiv Detail & Related papers (2020-03-16T12:12:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.