Time and Cost-Efficient Bathymetric Mapping System using Sparse Point
Cloud Generation and Automatic Object Detection
- URL: http://arxiv.org/abs/2210.10263v1
- Date: Wed, 19 Oct 2022 02:58:08 GMT
- Title: Time and Cost-Efficient Bathymetric Mapping System using Sparse Point
Cloud Generation and Automatic Object Detection
- Authors: Andres Pulido, Ruoyao Qin, Antonio Diaz, Andrew Ortega, Peter Ifju,
Jaejeong Shin
- Abstract summary: Side-scan sonar sensors are available in inexpensive cost ranges, especially in fish-finders.
Extracting 3D information from side-scan sonar imagery is a difficult task because of its low signal-to-noise ratio.
This paper introduces an efficient algorithm that generates a sparse 3D point cloud from side-scan sonar images.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Generating 3D point cloud (PC) data from noisy sonar measurements is a
problem that has potential applications for bathymetry mapping, artificial
object inspection, mapping of aquatic plants and fauna as well as underwater
navigation and localization of vehicles such as submarines. Side-scan sonar
sensors are available in inexpensive cost ranges, especially in fish-finders,
where the transducers are usually mounted to the bottom of a boat and can
approach shallower depths than the ones attached to an Uncrewed Underwater
Vehicle (UUV) can. However, extracting 3D information from side-scan sonar
imagery is a difficult task because of its low signal-to-noise ratio and
missing angle and depth information in the imagery. Since most algorithms that
generate a 3D point cloud from side-scan sonar imagery use Shape from Shading
(SFS) techniques, extracting 3D information is especially difficult when the
seafloor is smooth, is slowly changing in depth, or does not have identifiable
objects that make acoustic shadows. This paper introduces an efficient
algorithm that generates a sparse 3D point cloud from side-scan sonar images.
This computation is done in a computationally efficient manner by leveraging
the geometry of the first sonar return combined with known positions provided
by GPS and down-scan sonar depth measurement at each data point. Additionally,
this paper implements another algorithm that uses a Convolutional Neural
Network (CNN) using transfer learning to perform object detection on side-scan
sonar images collected in real life and generated with a simulation. The
algorithm was tested on both real and synthetic images to show reasonably
accurate anomaly detection and classification.
Related papers
- MVSDet: Multi-View Indoor 3D Object Detection via Efficient Plane Sweeps [51.44887282336391]
Key challenge of multi-view indoor 3D object detection is to infer accurate geometry information from images for precise 3D detection.
Previous method relies on NeRF for geometry reasoning.
We propose MVSDet which utilizes plane sweep for geometry-aware 3D object detection.
arXiv Detail & Related papers (2024-10-28T21:58:41Z) - Aug3D-RPN: Improving Monocular 3D Object Detection by Synthetic Images
with Virtual Depth [64.29043589521308]
We propose a rendering module to augment the training data by synthesizing images with virtual-depths.
The rendering module takes as input the RGB image and its corresponding sparse depth image, outputs a variety of photo-realistic synthetic images.
Besides, we introduce an auxiliary module to improve the detection model by jointly optimizing it through a depth estimation task.
arXiv Detail & Related papers (2021-07-28T11:00:47Z) - VR3Dense: Voxel Representation Learning for 3D Object Detection and
Monocular Dense Depth Reconstruction [0.951828574518325]
We introduce a method for jointly training 3D object detection and monocular dense depth reconstruction neural networks.
It takes as inputs, a LiDAR point-cloud, and a single RGB image during inference and produces object pose predictions as well as a densely reconstructed depth map.
While our object detection is trained in a supervised manner, the depth prediction network is trained with both self-supervised and supervised loss functions.
arXiv Detail & Related papers (2021-04-13T04:25:54Z) - PLUME: Efficient 3D Object Detection from Stereo Images [95.31278688164646]
Existing methods tackle the problem in two steps: first depth estimation is performed, a pseudo LiDAR point cloud representation is computed from the depth estimates, and then object detection is performed in 3D space.
We propose a model that unifies these two tasks in the same metric space.
Our approach achieves state-of-the-art performance on the challenging KITTI benchmark, with significantly reduced inference time compared with existing methods.
arXiv Detail & Related papers (2021-01-17T05:11:38Z) - Towards Dense People Detection with Deep Learning and Depth images [9.376814409561726]
This paper proposes a DNN-based system that detects multiple people from a single depth image.
Our neural network processes a depth image and outputs a likelihood map in image coordinates.
We show this strategy to be effective, producing networks that generalize to work with scenes different from those used during training.
arXiv Detail & Related papers (2020-07-14T16:43:02Z) - Expandable YOLO: 3D Object Detection from RGB-D Images [64.14512458954344]
This paper aims at constructing a light-weight object detector that inputs a depth and a color image from a stereo camera.
By extending the network architecture of YOLOv3 to 3D in the middle, it is possible to output in the depth direction.
Intersection over Uninon (IoU) in 3D space is introduced to confirm the accuracy of region extraction results.
arXiv Detail & Related papers (2020-06-26T07:32:30Z) - Stereo RGB and Deeper LIDAR Based Network for 3D Object Detection [40.34710686994996]
3D object detection has become an emerging task in autonomous driving scenarios.
Previous works process 3D point clouds using either projection-based or voxel-based models.
We propose the Stereo RGB and Deeper LIDAR framework which can utilize semantic and spatial information simultaneously.
arXiv Detail & Related papers (2020-06-09T11:19:24Z) - 3D Object Detection Method Based on YOLO and K-Means for Image and Point
Clouds [1.9458156037869139]
Lidar based 3D object detection and classification tasks are essential for autonomous driving.
This paper proposes a 3D object detection method based on point cloud and image.
arXiv Detail & Related papers (2020-04-21T04:32:36Z) - DOPS: Learning to Detect 3D Objects and Predict their 3D Shapes [54.239416488865565]
We propose a fast single-stage 3D object detection method for LIDAR data.
The core novelty of our method is a fast, single-pass architecture that both detects objects in 3D and estimates their shapes.
We find that our proposed method achieves state-of-the-art results by 5% on object detection in ScanNet scenes, and it gets top results by 3.4% in the Open dataset.
arXiv Detail & Related papers (2020-04-02T17:48:50Z) - DSGN: Deep Stereo Geometry Network for 3D Object Detection [79.16397166985706]
There is a large performance gap between image-based and LiDAR-based 3D object detectors.
Our method, called Deep Stereo Geometry Network (DSGN), significantly reduces this gap.
For the first time, we provide a simple and effective one-stage stereo-based 3D detection pipeline.
arXiv Detail & Related papers (2020-01-10T11:44:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.