3D Object Detection Method Based on YOLO and K-Means for Image and Point
Clouds
- URL: http://arxiv.org/abs/2005.02132v1
- Date: Tue, 21 Apr 2020 04:32:36 GMT
- Title: 3D Object Detection Method Based on YOLO and K-Means for Image and Point
Clouds
- Authors: Xuanyu Yin, Yoko Sasaki, Weimin Wang, Kentaro Shimizu
- Abstract summary: Lidar based 3D object detection and classification tasks are essential for autonomous driving.
This paper proposes a 3D object detection method based on point cloud and image.
- Score: 1.9458156037869139
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Lidar based 3D object detection and classification tasks are essential for
autonomous driving(AD). A lidar sensor can provide the 3D point cloud data
reconstruction of the surrounding environment. However, real time detection in
3D point clouds still needs a strong algorithmic. This paper proposes a 3D
object detection method based on point cloud and image which consists of there
parts.(1)Lidar-camera calibration and undistorted image transformation.
(2)YOLO-based detection and PointCloud extraction, (3)K-means based point cloud
segmentation and detection experiment test and evaluation in depth image. In
our research, camera can capture the image to make the Real-time 2D object
detection by using YOLO, we transfer the bounding box to node whose function is
making 3d object detection on point cloud data from Lidar. By comparing whether
2D coordinate transferred from the 3D point is in the object bounding box or
not can achieve High-speed 3D object recognition function in GPU. The accuracy
and precision get imporved after k-means clustering in point cloud. The speed
of our detection method is a advantage faster than PointNet.
Related papers
- Sparse2Dense: Learning to Densify 3D Features for 3D Object Detection [85.08249413137558]
LiDAR-produced point clouds are the major source for most state-of-the-art 3D object detectors.
Small, distant, and incomplete objects with sparse or few points are often hard to detect.
We present Sparse2Dense, a new framework to efficiently boost 3D detection performance by learning to densify point clouds in latent space.
arXiv Detail & Related papers (2022-11-23T16:01:06Z) - Boosting 3D Object Detection via Object-Focused Image Fusion [33.616129400275156]
We present DeMF, a method to fuse image information into point features.
We evaluate our method on the challenging SUN RGB-D dataset.
arXiv Detail & Related papers (2022-07-21T16:32:05Z) - Image-to-Lidar Self-Supervised Distillation for Autonomous Driving Data [80.14669385741202]
We propose a self-supervised pre-training method for 3D perception models tailored to autonomous driving data.
We leverage the availability of synchronized and calibrated image and Lidar sensors in autonomous driving setups.
Our method does not require any point cloud nor image annotations.
arXiv Detail & Related papers (2022-03-30T12:40:30Z) - VPFNet: Improving 3D Object Detection with Virtual Point based LiDAR and
Stereo Data Fusion [62.24001258298076]
VPFNet is a new architecture that cleverly aligns and aggregates the point cloud and image data at the virtual' points.
Our VPFNet achieves 83.21% moderate 3D AP and 91.86% moderate BEV AP on the KITTI test set, ranking the 1st since May 21th, 2021.
arXiv Detail & Related papers (2021-11-29T08:51:20Z) - Anchor-free 3D Single Stage Detector with Mask-Guided Attention for
Point Cloud [79.39041453836793]
We develop a novel single-stage 3D detector for point clouds in an anchor-free manner.
We overcome this by converting the voxel-based sparse 3D feature volumes into the sparse 2D feature maps.
We propose an IoU-based detection confidence re-calibration scheme to improve the correlation between the detection confidence score and the accuracy of the bounding box regression.
arXiv Detail & Related papers (2021-08-08T13:42:13Z) - Group-Free 3D Object Detection via Transformers [26.040378025818416]
We present a simple yet effective method for directly detecting 3D objects from the 3D point cloud.
Our method computes the feature of an object from all the points in the point cloud with the help of an attention mechanism in the Transformers citevaswaniattention.
With few bells and whistles, the proposed method achieves state-of-the-art 3D object detection performance on two widely used benchmarks, ScanNet V2 and SUN RGB-D.
arXiv Detail & Related papers (2021-04-01T17:59:36Z) - RoIFusion: 3D Object Detection from LiDAR and Vision [7.878027048763662]
We propose a novel fusion algorithm by projecting a set of 3D Region of Interests (RoIs) from the point clouds to the 2D RoIs of the corresponding the images.
Our approach achieves state-of-the-art performance on the KITTI 3D object detection challenging benchmark.
arXiv Detail & Related papers (2020-09-09T20:23:27Z) - YOLO and K-Means Based 3D Object Detection Method on Image and Point
Cloud [1.9458156037869139]
Lidar based 3D object detection and classification tasks are essential for automated driving.
This paper consists of three parts.
Camera can capture the image to make the Real-time 2D Object Detection by using YOLO.
By comparing whether 2D coordinate transferred from the 3D point is in the object bounding box or not, and doing a k-means clustering can achieve High-speed 3D object recognition function in GPU.
arXiv Detail & Related papers (2020-04-21T03:08:46Z) - DOPS: Learning to Detect 3D Objects and Predict their 3D Shapes [54.239416488865565]
We propose a fast single-stage 3D object detection method for LIDAR data.
The core novelty of our method is a fast, single-pass architecture that both detects objects in 3D and estimates their shapes.
We find that our proposed method achieves state-of-the-art results by 5% on object detection in ScanNet scenes, and it gets top results by 3.4% in the Open dataset.
arXiv Detail & Related papers (2020-04-02T17:48:50Z) - Boundary-Aware Dense Feature Indicator for Single-Stage 3D Object
Detection from Point Clouds [32.916690488130506]
We propose a universal module that helps 3D detectors focus on the densest region of the point clouds in a boundary-aware manner.
Experiments on KITTI dataset show that DENFI improves the performance of the baseline single-stage detector remarkably.
arXiv Detail & Related papers (2020-04-01T01:21:23Z) - ImVoteNet: Boosting 3D Object Detection in Point Clouds with Image Votes [93.82668222075128]
We propose a 3D detection architecture called ImVoteNet for RGB-D scenes.
ImVoteNet is based on fusing 2D votes in images and 3D votes in point clouds.
We validate our model on the challenging SUN RGB-D dataset.
arXiv Detail & Related papers (2020-01-29T05:09:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.