R-C-P Method: An Autonomous Volume Calculation Method Using Image
Processing and Machine Vision
- URL: http://arxiv.org/abs/2308.10058v2
- Date: Sun, 4 Feb 2024 01:43:07 GMT
- Title: R-C-P Method: An Autonomous Volume Calculation Method Using Image
Processing and Machine Vision
- Authors: MA Muktadir, Sydney Parker, Sun Yi
- Abstract summary: Two cameras were used to measure the dimensions of a rectangular object in real-time.
The R-C-P method is developed using image processing and edge detection.
In addition to the surface areas, the R-C-P method also detects discontinuous edges or volumes.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Machine vision and image processing are often used with sensors for situation
awareness in autonomous systems, from industrial robots to self-driving cars.
The 3D depth sensors, such as LiDAR (Light Detection and Ranging), Radar, are
great invention for autonomous systems. Due to the complexity of the setup,
LiDAR may not be suitable for some operational environments, for example, a
space environment. This study was motivated by a desire to get real-time
volumetric and change information with multiple 2D cameras instead of a depth
camera. Two cameras were used to measure the dimensions of a rectangular object
in real-time. The R-C-P (row-column-pixel) method is developed using image
processing and edge detection. In addition to the surface areas, the R-C-P
method also detects discontinuous edges or volumes. Lastly, experimental work
is presented for illustration of the R-C-P method, which provides the equations
for calculating surface area dimensions. Using the equations with given
distance information between the object and the camera, the vision system
provides the dimensions of actual objects.
Related papers
- Multi-Modal Dataset Acquisition for Photometrically Challenging Object [56.30027922063559]
This paper addresses the limitations of current datasets for 3D vision tasks in terms of accuracy, size, realism, and suitable imaging modalities for photometrically challenging objects.
We propose a novel annotation and acquisition pipeline that enhances existing 3D perception and 6D object pose datasets.
arXiv Detail & Related papers (2023-08-21T10:38:32Z) - 3D Data Augmentation for Driving Scenes on Camera [50.41413053812315]
We propose a 3D data augmentation approach termed Drive-3DAug, aiming at augmenting the driving scenes on camera in the 3D space.
We first utilize Neural Radiance Field (NeRF) to reconstruct the 3D models of background and foreground objects.
Then, augmented driving scenes can be obtained by placing the 3D objects with adapted location and orientation at the pre-defined valid region of backgrounds.
arXiv Detail & Related papers (2023-03-18T05:51:05Z) - Vision-Based Environmental Perception for Autonomous Driving [4.138893879750758]
Visual perception plays an important role in autonomous driving.
Recent development of deep learning-based method has better reliability and processing speed.
Monocular camera uses image data from a single viewpoint to estimate object depth.
Simultaneous Location and Mapping (SLAM) can establish a model of the road environment.
arXiv Detail & Related papers (2022-12-22T01:59:58Z) - Image-to-Lidar Self-Supervised Distillation for Autonomous Driving Data [80.14669385741202]
We propose a self-supervised pre-training method for 3D perception models tailored to autonomous driving data.
We leverage the availability of synchronized and calibrated image and Lidar sensors in autonomous driving setups.
Our method does not require any point cloud nor image annotations.
arXiv Detail & Related papers (2022-03-30T12:40:30Z) - Disentangling and Vectorization: A 3D Visual Perception Approach for
Autonomous Driving Based on Surround-View Fisheye Cameras [3.485767750936058]
Multidimensional Vector is proposed to include the utilizable information generated in different dimensions and stages.
The experiments of real fisheye images demonstrate that our solution achieves state-of-the-art accuracy while being real-time in practice.
arXiv Detail & Related papers (2021-07-19T13:24:21Z) - High-level camera-LiDAR fusion for 3D object detection with machine
learning [0.0]
This paper tackles the 3D object detection problem, which is of vital importance for applications such as autonomous driving.
It uses a Machine Learning pipeline on a combination of monocular camera and LiDAR data to detect vehicles in the surrounding 3D space of a moving platform.
Our results demonstrate an efficient and accurate inference on a validation set, achieving an overall accuracy of 87.1%.
arXiv Detail & Related papers (2021-05-24T01:57:34Z) - Supervised Training of Dense Object Nets using Optimal Descriptors for
Industrial Robotic Applications [57.87136703404356]
Dense Object Nets (DONs) by Florence, Manuelli and Tedrake introduced dense object descriptors as a novel visual object representation for the robotics community.
In this paper we show that given a 3D model of an object, we can generate its descriptor space image, which allows for supervised training of DONs.
We compare the training methods on generating 6D grasps for industrial objects and show that our novel supervised training approach improves the pick-and-place performance in industry-relevant tasks.
arXiv Detail & Related papers (2021-02-16T11:40:12Z) - PLUME: Efficient 3D Object Detection from Stereo Images [95.31278688164646]
Existing methods tackle the problem in two steps: first depth estimation is performed, a pseudo LiDAR point cloud representation is computed from the depth estimates, and then object detection is performed in 3D space.
We propose a model that unifies these two tasks in the same metric space.
Our approach achieves state-of-the-art performance on the challenging KITTI benchmark, with significantly reduced inference time compared with existing methods.
arXiv Detail & Related papers (2021-01-17T05:11:38Z) - Lightweight Multi-View 3D Pose Estimation through Camera-Disentangled
Representation [57.11299763566534]
We present a solution to recover 3D pose from multi-view images captured with spatially calibrated cameras.
We exploit 3D geometry to fuse input images into a unified latent representation of pose, which is disentangled from camera view-points.
Our architecture then conditions the learned representation on camera projection operators to produce accurate per-view 2d detections.
arXiv Detail & Related papers (2020-04-05T12:52:29Z) - Automatic Estimation of Sphere Centers from Images of Calibrated Cameras [11.816942730023886]
This paper deals with the automatic detection of ellipses in camera images, as well as to estimate the 3D position of the spheres corresponding to the detected 2D ellipses.
We propose two novel methods to (i) detect an ellipse in camera images and (ii) estimate the spatial location of the corresponding sphere if its size is known.
They are applied for calibrating the sensor system of autonomous cars equipped with digital cameras, depth sensors and LiDAR devices.
arXiv Detail & Related papers (2020-02-24T13:12:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.