Semantics for Robotic Mapping, Perception and Interaction: A Survey
- URL: http://arxiv.org/abs/2101.00443v1
- Date: Sat, 2 Jan 2021 12:34:39 GMT
- Title: Semantics for Robotic Mapping, Perception and Interaction: A Survey
- Authors: Sourav Garg, Niko S\"underhauf, Feras Dayoub, Douglas Morrison,
Akansel Cosgun, Gustavo Carneiro, Qi Wu, Tat-Jun Chin, Ian Reid, Stephen
Gould, Peter Corke, Michael Milford
- Abstract summary: Study of understanding dictates what does the world "mean" to a robot.
With humans and robots increasingly operating in the same world, the prospects of human-robot interaction also bring semantics into the picture.
Driven by need, as well as by enablers like increasing availability of training data and computational resources, semantics is a rapidly growing research area in robotics.
- Score: 93.93587844202534
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: For robots to navigate and interact more richly with the world around them,
they will likely require a deeper understanding of the world in which they
operate. In robotics and related research fields, the study of understanding is
often referred to as semantics, which dictates what does the world "mean" to a
robot, and is strongly tied to the question of how to represent that meaning.
With humans and robots increasingly operating in the same world, the prospects
of human-robot interaction also bring semantics and ontology of natural
language into the picture. Driven by need, as well as by enablers like
increasing availability of training data and computational resources, semantics
is a rapidly growing research area in robotics. The field has received
significant attention in the research literature to date, but most reviews and
surveys have focused on particular aspects of the topic: the technical research
issues regarding its use in specific robotic topics like mapping or
segmentation, or its relevance to one particular application domain like
autonomous driving. A new treatment is therefore required, and is also timely
because so much relevant research has occurred since many of the key surveys
were published. This survey therefore provides an overarching snapshot of where
semantics in robotics stands today. We establish a taxonomy for semantics
research in or relevant to robotics, split into four broad categories of
activity, in which semantics are extracted, used, or both. Within these broad
categories we survey dozens of major topics including fundamentals from the
computer vision field and key robotics research areas utilizing semantics,
including mapping, navigation and interaction with the world. The survey also
covers key practical considerations, including enablers like increased data
availability and improved computational hardware, and major application areas
where...
Related papers
- A Survey of Embodied Learning for Object-Centric Robotic Manipulation [27.569063968870868]
Embodied learning for object-centric robotic manipulation is a rapidly developing and challenging area in AI.
Unlike data-driven machine learning methods, embodied learning focuses on robot learning through physical interaction with the environment.
arXiv Detail & Related papers (2024-08-21T11:32:09Z) - Toward General-Purpose Robots via Foundation Models: A Survey and Meta-Analysis [82.59451639072073]
General-purpose robots operate seamlessly in any environment, with any object, and utilize various skills to complete diverse tasks.
As a community, we have been constraining most robotic systems by designing them for specific tasks, training them on specific datasets, and deploying them within specific environments.
Motivated by the impressive open-set performance and content generation capabilities of web-scale, large-capacity pre-trained models, we devote this survey to exploring how foundation models can be applied to general-purpose robotics.
arXiv Detail & Related papers (2023-12-14T10:02:55Z) - Open-World Object Manipulation using Pre-trained Vision-Language Models [72.87306011500084]
For robots to follow instructions from people, they must be able to connect the rich semantic information in human vocabulary.
We develop a simple approach, which leverages a pre-trained vision-language model to extract object-identifying information.
In a variety of experiments on a real mobile manipulator, we find that MOO generalizes zero-shot to a wide range of novel object categories and environments.
arXiv Detail & Related papers (2023-03-02T01:55:10Z) - Deep Learning for Human Parsing: A Survey [54.812353922568995]
We provide an analysis of state-of-the-art human parsing methods, covering a broad spectrum of pioneering works for semantic human parsing.
We introduce five insightful categories: (1) structure-driven architectures exploit the relationship of different human parts and the inherent hierarchical structure of a human body, (2) graph-based networks capture the global information to achieve an efficient and complete human body analysis, (3) context-aware networks explore useful contexts across all pixel to characterize a pixel of the corresponding class, and (4) LSTM-based methods can combine short-distance and long-distance spatial dependencies to better exploit abundant local and global contexts.
arXiv Detail & Related papers (2023-01-29T10:54:56Z) - Semantic-Aware Environment Perception for Mobile Human-Robot Interaction [2.309914459672557]
We present a vision-based system for mobile robots to enable a semantic-aware environment without additional a-priori knowledge.
We deploy our system on a mobile humanoid robot that enables us to test our methods in real-world applications.
arXiv Detail & Related papers (2022-11-07T08:49:45Z) - Augmented Reality and Robotics: A Survey and Taxonomy for AR-enhanced
Human-Robot Interaction and Robotic Interfaces [21.72715373737208]
Augmented and mixed reality (AR/MR) have emerged as a new way to enhance human-robot interaction.
This paper contributes to a taxonomy of augmented reality and robotics based on a survey of 460 research papers.
arXiv Detail & Related papers (2022-03-07T10:22:59Z) - Human-Robot Collaboration and Machine Learning: A Systematic Review of
Recent Research [69.48907856390834]
Human-robot collaboration (HRC) is the approach that explores the interaction between a human and a robot.
This paper proposes a thorough literature review of the use of machine learning techniques in the context of HRC.
arXiv Detail & Related papers (2021-10-14T15:14:33Z) - A Survey on Human-aware Robot Navigation [22.926878248788558]
This paper is concerned with the navigation aspect of a socially-compliant robot.
It provides a survey of existing solutions for the relevant areas of research as well as an outlook on possible future directions.
arXiv Detail & Related papers (2021-06-22T10:09:25Z) - Spoken Language Interaction with Robots: Research Issues and
Recommendations, Report from the NSF Future Directions Workshop [0.819605661841562]
Meeting human needs requires addressing new challenges in speech technology and user experience design.
More powerful adaptation methods are needed, without extensive re-engineering or the collection of massive training data.
Since robots operate in real time, their speech and language processing components must also.
arXiv Detail & Related papers (2020-11-11T03:45:34Z) - Task-relevant Representation Learning for Networked Robotic Perception [74.0215744125845]
This paper presents an algorithm to learn task-relevant representations of sensory data that are co-designed with a pre-trained robotic perception model's ultimate objective.
Our algorithm aggressively compresses robotic sensory data by up to 11x more than competing methods.
arXiv Detail & Related papers (2020-11-06T07:39:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.