Semantic Scene Segmentation for Robotics Applications
- URL: http://arxiv.org/abs/2108.11128v1
- Date: Wed, 25 Aug 2021 08:55:20 GMT
- Title: Semantic Scene Segmentation for Robotics Applications
- Authors: Maria Tzelepi and Anastasios Tefas
- Abstract summary: We investigate the behavior of the most successful semantic scene segmentation models, in terms of deployment (inference) speed, under various setups.
The target of this work is to provide a comparative study of current state-of-the-art segmentation models so as to select the most compliant with the robotics applications requirements.
- Score: 51.66271681532262
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Semantic scene segmentation plays a critical role in a wide range of robotics
applications, e.g., autonomous navigation. These applications are accompanied
by specific computational restrictions, e.g., operation on low-power GPUs, at
sufficient speed, and also for high-resolution input. Existing state-of-the-art
segmentation models provide evaluation results under different setups and
mainly considering high-power GPUs. In this paper, we investigate the behavior
of the most successful semantic scene segmentation models, in terms of
deployment (inference) speed, under various setups (GPUs, input sizes, etc.) in
the context of robotics applications. The target of this work is to provide a
comparative study of current state-of-the-art segmentation models so as to
select the most compliant with the robotics applications requirements.
Related papers
- Are We Ready for Real-Time LiDAR Semantic Segmentation in Autonomous Driving? [42.348499880894686]
Scene semantic segmentation can be achieved by directly integrating 3D spatial data with specialized deep neural networks.
We investigate various 3D semantic segmentation methodologies and analyze their performance and capabilities for resource-constrained inference on embedded NVIDIA Jetson platforms.
arXiv Detail & Related papers (2024-10-10T20:47:33Z) - Show and Grasp: Few-shot Semantic Segmentation for Robot Grasping through Zero-shot Foundation Models [5.792788640304759]
The ability of a robot to pick an object, known as robot grasping, is crucial for several applications.
In such tasks, selecting the right target to pick is as essential as inferring a correct configuration of the gripper.
A common solution to this problem relies on semantic segmentation models, which often show poor generalization to unseen objects.
arXiv Detail & Related papers (2024-04-19T08:58:52Z) - Instruct2Act: Mapping Multi-modality Instructions to Robotic Actions
with Large Language Model [63.66204449776262]
Instruct2Act is a framework that maps multi-modal instructions to sequential actions for robotic manipulation tasks.
Our approach is adjustable and flexible in accommodating various instruction modalities and input types.
Our zero-shot method outperformed many state-of-the-art learning-based policies in several tasks.
arXiv Detail & Related papers (2023-05-18T17:59:49Z) - A Threefold Review on Deep Semantic Segmentation: Efficiency-oriented,
Temporal and Depth-aware design [77.34726150561087]
We conduct a survey on the most relevant and recent advances in Deep Semantic in the context of vision for autonomous vehicles.
Our main objective is to provide a comprehensive discussion on the main methods, advantages, limitations, results and challenges faced from each perspective.
arXiv Detail & Related papers (2023-03-08T01:29:55Z) - On Efficient Real-Time Semantic Segmentation: A Survey [12.404169549562523]
We take a look at the works that aim to address this misalignment with more compact and efficient models capable of deployment on low-memory embedded systems.
We evaluate the inference speed of the discussed models under consistent hardware and software setups.
Our experimental results demonstrate that many works are capable of real-time performance on resource-constrained hardware, while illustrating the consistent trade-off between latency and accuracy.
arXiv Detail & Related papers (2022-06-17T08:00:27Z) - OG-SGG: Ontology-Guided Scene Graph Generation. A Case Study in Transfer
Learning for Telepresence Robotics [124.08684545010664]
Scene graph generation from images is a task of great interest to applications such as robotics.
We propose an initial approximation to a framework called Ontology-Guided Scene Graph Generation (OG-SGG)
arXiv Detail & Related papers (2022-02-21T13:23:15Z) - MetaGraspNet: A Large-Scale Benchmark Dataset for Vision-driven Robotic
Grasping via Physics-based Metaverse Synthesis [78.26022688167133]
We present a large-scale benchmark dataset for vision-driven robotic grasping via physics-based metaverse synthesis.
The proposed dataset contains 100,000 images and 25 different object types.
We also propose a new layout-weighted performance metric alongside the dataset for evaluating object detection and segmentation performance.
arXiv Detail & Related papers (2021-12-29T17:23:24Z) - Efficient Multi-Organ Segmentation Using SpatialConfiguration-Net with
Low GPU Memory Requirements [8.967700713755281]
In this work, we employ a multi-organ segmentation model based on the SpatialConfiguration-Net (SCN)
We modified the architecture of the segmentation model to reduce its memory footprint without drastically impacting the quality of the predictions.
Lastly, we implemented a minimal inference script for which we optimized both, execution time and required GPU memory.
arXiv Detail & Related papers (2021-11-26T17:47:10Z) - SideInfNet: A Deep Neural Network for Semi-Automatic Semantic
Segmentation with Side Information [83.03179580646324]
This paper proposes a novel deep neural network architecture, namely SideInfNet.
It integrates features learnt from images with side information extracted from user annotations.
To evaluate our method, we applied the proposed network to three semantic segmentation tasks and conducted extensive experiments on benchmark datasets.
arXiv Detail & Related papers (2020-02-07T06:10:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.