COVERED, CollabOratiVE Robot Environment Dataset for 3D Semantic
segmentation
- URL: http://arxiv.org/abs/2302.12656v2
- Date: Tue, 4 Apr 2023 09:06:52 GMT
- Title: COVERED, CollabOratiVE Robot Environment Dataset for 3D Semantic
segmentation
- Authors: Charith Munasinghe, Fatemeh Mohammadi Amin, Davide Scaramuzza, Hans
Wernher van de Venn
- Abstract summary: This work develops a new dataset specifically designed for this use case, named "COVERED"
We provide a benchmark of current state-of-the-art (SOTA) algorithm performance on the dataset and demonstrate a real-time semantic segmentation of a collaborative robot workspace using a multi-LiDAR system.
Our perception pipeline achieves 20Hz throughput with a prediction point accuracy of $>$96% and $>$92% mean intersection over union (mIOU) while maintaining an 8Hz throughput.
- Score: 39.64058995273062
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Safe human-robot collaboration (HRC) has recently gained a lot of interest
with the emerging Industry 5.0 paradigm. Conventional robots are being replaced
with more intelligent and flexible collaborative robots (cobots). Safe and
efficient collaboration between cobots and humans largely relies on the cobot's
comprehensive semantic understanding of the dynamic surrounding of industrial
environments. Despite the importance of semantic understanding for such
applications, 3D semantic segmentation of collaborative robot workspaces lacks
sufficient research and dedicated datasets. The performance limitation caused
by insufficient datasets is called 'data hunger' problem. To overcome this
current limitation, this work develops a new dataset specifically designed for
this use case, named "COVERED", which includes point-wise annotated point
clouds of a robotic cell. Lastly, we also provide a benchmark of current
state-of-the-art (SOTA) algorithm performance on the dataset and demonstrate a
real-time semantic segmentation of a collaborative robot workspace using a
multi-LiDAR system. The promising results from using the trained Deep Networks
on a real-time dynamically changing situation shows that we are on the right
track. Our perception pipeline achieves 20Hz throughput with a prediction point
accuracy of $>$96\% and $>$92\% mean intersection over union (mIOU) while
maintaining an 8Hz throughput.
Related papers
- The Collection of a Human Robot Collaboration Dataset for Cooperative Assembly in Glovebox Environments [2.30069810310356]
Industry 4.0 introduced AI as a transformative solution for modernizing manufacturing processes. Its successor, Industry 5.0, envisions humans as collaborators and experts guiding these AI-driven solutions.
New techniques require algorithms capable of safe, real-time identification of human positions in a scene, particularly their hands, during collaborative assembly.
This dataset provides 1200 challenging examples to build applications toward hand and glove segmentation in industrial human collaboration scenarios.
arXiv Detail & Related papers (2024-07-19T19:56:53Z) - LPAC: Learnable Perception-Action-Communication Loops with Applications
to Coverage Control [80.86089324742024]
We propose a learnable Perception-Action-Communication (LPAC) architecture for the problem.
CNN processes localized perception; a graph neural network (GNN) facilitates robot communications.
Evaluations show that the LPAC models outperform standard decentralized and centralized coverage control algorithms.
arXiv Detail & Related papers (2024-01-10T00:08:00Z) - Care3D: An Active 3D Object Detection Dataset of Real Robotic-Care
Environments [52.425280825457385]
This paper introduces an annotated dataset of real environments.
The captured environments represent areas which are already in use in the field of robotic health care research.
We also provide ground truth data within one room, for assessing SLAM algorithms running directly on a health care robot.
arXiv Detail & Related papers (2023-10-09T10:35:37Z) - HabitatDyn Dataset: Dynamic Object Detection to Kinematics Estimation [16.36110033895749]
We propose the dataset HabitatDyn, which contains both synthetic RGB videos, semantic labels, and depth information, as well as kinetics information.
HabitatDyn was created from the perspective of a mobile robot with a moving camera, and contains 30 scenes featuring six different types of moving objects with varying velocities.
arXiv Detail & Related papers (2023-04-21T09:57:35Z) - Model Predictive Control for Fluid Human-to-Robot Handovers [50.72520769938633]
Planning motions that take human comfort into account is not a part of the human-robot handover process.
We propose to generate smooth motions via an efficient model-predictive control framework.
We conduct human-to-robot handover experiments on a diverse set of objects with several users.
arXiv Detail & Related papers (2022-03-31T23:08:20Z) - Regularized Deep Signed Distance Fields for Reactive Motion Generation [30.792481441975585]
Distance-based constraints are fundamental for enabling robots to plan their actions and act safely.
We propose Regularized Deep Signed Distance Fields (ReDSDF), a single neural implicit function that can compute smooth distance fields at any scale.
We demonstrate the effectiveness of our approach in representative simulated tasks for whole-body control (WBC) and safe Human-Robot Interaction (HRI) in shared workspaces.
arXiv Detail & Related papers (2022-03-09T14:21:32Z) - HARPS: An Online POMDP Framework for Human-Assisted Robotic Planning and
Sensing [1.3678064890824186]
The Human Assisted Robotic Planning and Sensing (HARPS) framework is presented for active semantic sensing and planning in human-robot teams.
This approach lets humans opportunistically impose model structure and extend the range of semantic soft data in uncertain environments.
Simulations of a UAV-enabled target search application in a large-scale partially structured environment show significant improvements in time and belief state estimates.
arXiv Detail & Related papers (2021-10-20T00:41:57Z) - Cross-Modal Analysis of Human Detection for Robotics: An Industrial Case
Study [7.844709223688293]
We conduct a systematic cross-modal analysis of sensor-algorithm combinations typically used in robotics.
We compare the performance of state-of-the-art person detectors for 2D range data, 3D lidar, and RGB-D data.
We extend a strong image-based RGB-D detector to provide cross-modal supervision for lidar detectors in the form of weak 3D bounding box labels.
arXiv Detail & Related papers (2021-08-03T13:33:37Z) - SABER: Data-Driven Motion Planner for Autonomously Navigating
Heterogeneous Robots [112.2491765424719]
We present an end-to-end online motion planning framework that uses a data-driven approach to navigate a heterogeneous robot team towards a global goal.
We use model predictive control (SMPC) to calculate control inputs that satisfy robot dynamics, and consider uncertainty during obstacle avoidance with chance constraints.
recurrent neural networks are used to provide a quick estimate of future state uncertainty considered in the SMPC finite-time horizon solution.
A Deep Q-learning agent is employed to serve as a high-level path planner, providing the SMPC with target positions that move the robots towards a desired global goal.
arXiv Detail & Related papers (2021-08-03T02:56:21Z) - Where is my hand? Deep hand segmentation for visual self-recognition in
humanoid robots [129.46920552019247]
We propose the use of a Convolution Neural Network (CNN) to segment the robot hand from an image in an egocentric view.
We fine-tuned the Mask-RCNN network for the specific task of segmenting the hand of the humanoid robot Vizzy.
arXiv Detail & Related papers (2021-02-09T10:34:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.