Who's in Charge? Roles and Responsibilities of Decision-Making
Components in Conversational Robots
- URL: http://arxiv.org/abs/2303.08470v1
- Date: Wed, 15 Mar 2023 09:18:32 GMT
- Title: Who's in Charge? Roles and Responsibilities of Decision-Making
Components in Conversational Robots
- Authors: Pierre Lison and Casey Kennington
- Abstract summary: We reflect on the organization of decision modules in human-robot interaction platforms.
We show that most practical HRI architectures tend to be either robot-centric or dialogue-centric.
We contend that architectures placing action managers'' and interaction managers'' on an equal footing may provide the best path forward for future human-robot interaction systems.
- Score: 3.7311680121118345
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Software architectures for conversational robots typically consist of
multiple modules, each designed for a particular processing task or
functionality. Some of these modules are developed for the purpose of making
decisions about the next action that the robot ought to perform in the current
context. Those actions may relate to physical movements, such as driving
forward or grasping an object, but may also correspond to communicative acts,
such as asking a question to the human user. In this position paper, we reflect
on the organization of those decision modules in human-robot interaction
platforms. We discuss the relative benefits and limitations of modular vs.
end-to-end architectures, and argue that, despite the increasing popularity of
end-to-end approaches, modular architectures remain preferable when developing
conversational robots designed to execute complex tasks in collaboration with
human users. We also show that most practical HRI architectures tend to be
either robot-centric or dialogue-centric, depending on where developers wish to
place the ``command center'' of their system. While those design choices may be
justified in some application domains, they also limit the robot's ability to
flexibly interleave physical movements and conversational behaviours. We
contend that architectures placing ``action managers'' and ``interaction
managers'' on an equal footing may provide the best path forward for future
human-robot interaction systems.
Related papers
- $π_0$: A Vision-Language-Action Flow Model for General Robot Control [77.32743739202543]
We propose a novel flow matching architecture built on top of a pre-trained vision-language model (VLM) to inherit Internet-scale semantic knowledge.
We evaluate our model in terms of its ability to perform tasks in zero shot after pre-training, follow language instructions from people, and its ability to acquire new skills via fine-tuning.
arXiv Detail & Related papers (2024-10-31T17:22:30Z) - Survey of Design Paradigms for Social Robots [10.618592615516901]
Social robots leverage multimodal communication, incorporating speech, facial expressions, and gestures to enhance user engagement and emotional support.
The understanding of design paradigms of social robots is obstructed by the complexity of the system and the necessity to tune it to a specific task.
This article provides a structured review of social robot design paradigms, categorizing them into cognitive architectures, role design models, linguistic models, communication flow, activity system models, and integrated design models.
arXiv Detail & Related papers (2024-07-30T05:22:31Z) - Nadine: An LLM-driven Intelligent Social Robot with Affective Capabilities and Human-like Memory [3.3906920519220054]
We describe our approach to developing an intelligent and robust social robotic system for the Nadine platform.
We achieve this by integrating Large Language Models (LLMs) and skilfully leveraging the powerful reasoning and instruction-following capabilities of these types of models.
This approach is novel compared to the current state-of-the-art LLM-based agents which do not implement human-like long-term memory or sophisticated emotional appraisal.
arXiv Detail & Related papers (2024-05-30T15:55:41Z) - RoboCodeX: Multimodal Code Generation for Robotic Behavior Synthesis [102.1876259853457]
We propose a tree-structured multimodal code generation framework for generalized robotic behavior synthesis, termed RoboCodeX.
RoboCodeX decomposes high-level human instructions into multiple object-centric manipulation units consisting of physical preferences such as affordance and safety constraints.
To further enhance the capability to map conceptual and perceptual understanding into control commands, a specialized multimodal reasoning dataset is collected for pre-training and an iterative self-updating methodology is introduced for supervised fine-tuning.
arXiv Detail & Related papers (2024-02-25T15:31:43Z) - RoboScript: Code Generation for Free-Form Manipulation Tasks across Real
and Simulation [77.41969287400977]
This paper presents textbfRobotScript, a platform for a deployable robot manipulation pipeline powered by code generation.
We also present a benchmark for a code generation benchmark for robot manipulation tasks in free-form natural language.
We demonstrate the adaptability of our code generation framework across multiple robot embodiments, including the Franka and UR5 robot arms.
arXiv Detail & Related papers (2024-02-22T15:12:00Z) - Model Predictive Control for Fluid Human-to-Robot Handovers [50.72520769938633]
Planning motions that take human comfort into account is not a part of the human-robot handover process.
We propose to generate smooth motions via an efficient model-predictive control framework.
We conduct human-to-robot handover experiments on a diverse set of objects with several users.
arXiv Detail & Related papers (2022-03-31T23:08:20Z) - Synthesis and Execution of Communicative Robotic Movements with
Generative Adversarial Networks [59.098560311521034]
We focus on how to transfer on two different robotic platforms the same kinematics modulation that humans adopt when manipulating delicate objects.
We choose to modulate the velocity profile adopted by the robots' end-effector, inspired by what humans do when transporting objects with different characteristics.
We exploit a novel Generative Adversarial Network architecture, trained with human kinematics examples, to generalize over them and generate new and meaningful velocity profiles.
arXiv Detail & Related papers (2022-03-29T15:03:05Z) - A ROS Architecture for Personalised HRI with a Bartender Social Robot [61.843727637976045]
BRILLO project has the overall goal of creating an autonomous robotic bartender that can interact with customers while accomplishing its bartending tasks.
We present the developed three-layers ROS architecture integrating a perception layer managing the processing of different social signals, a decision-making layer for handling multi-party interactions, and an execution layer controlling the behaviour of a complex robot composed of arms and a face.
arXiv Detail & Related papers (2022-03-13T11:33:06Z) - A MultiModal Social Robot Toward Personalized Emotion Interaction [1.2183405753834562]
This study demonstrates a multimodal human-robot interaction (HRI) framework with reinforcement learning to enhance the robotic interaction policy.
The goal is to apply this framework in social scenarios that can let the robots generate a more natural and engaging HRI framework.
arXiv Detail & Related papers (2021-10-08T00:35:44Z) - Behavior coordination for self-adaptive robots using constraint-based
configuration [0.0]
This paper presents an original algorithm to dynamically configure the control architecture of self-adaptive robots.
The algorithm uses a constraint-based configuration approach to decide which basic robot behaviors should be activated in response to both reactive and deliberative events.
The solution has been implemented as a software development tool called Behavior Coordinator CBC, which is based on ROS and open source.
arXiv Detail & Related papers (2021-03-24T12:09:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.