SOCRATES: Text-based Human Search and Approach using a Robot Dog
- URL: http://arxiv.org/abs/2302.05324v2
- Date: Sun, 18 Jun 2023 07:30:54 GMT
- Title: SOCRATES: Text-based Human Search and Approach using a Robot Dog
- Authors: Jeongeun Park, Jefferson Silveria, Matthew Pan, and Sungjoon Choi
- Abstract summary: We propose a SOCratic model for Robots Approaching humans based on TExt System (SOCRATES)
We first present a Human Search Socratic Model that connects large pre-trained models in the language domain to solve the downstream task.
Then, we propose a hybrid learning-based framework for generating target-cordial robotic motion to approach a person.
- Score: 6.168521568443759
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we propose a SOCratic model for Robots Approaching humans
based on TExt System (SOCRATES) focusing on the human search and approach based
on free-form textual description; the robot first searches for the target user,
then the robot proceeds to approach in a human-friendly manner. In particular,
textual descriptions are composed of appearance (e.g., wearing white shirts
with black hair) and location clues (e.g., is a student who works with robots).
We initially present a Human Search Socratic Model that connects large
pre-trained models in the language domain to solve the downstream task, which
is searching for the target person based on textual descriptions. Then, we
propose a hybrid learning-based framework for generating target-cordial robotic
motion to approach a person, consisting of a learning-from-demonstration module
and a knowledge distillation module. We validate the proposed searching module
via simulation using a virtual mobile robot as well as through real-world
experiments involving participants and the Boston Dynamics Spot robot.
Furthermore, we analyze the properties of the proposed approaching framework
with human participants based on the Robotic Social Attributes Scale (RoSAS)
Related papers
- HumanoidBench: Simulated Humanoid Benchmark for Whole-Body Locomotion and Manipulation [50.616995671367704]
We present a high-dimensional, simulated robot learning benchmark, HumanoidBench, featuring a humanoid robot equipped with dexterous hands.
Our findings reveal that state-of-the-art reinforcement learning algorithms struggle with most tasks, whereas a hierarchical learning approach achieves superior performance when supported by robust low-level policies.
arXiv Detail & Related papers (2024-03-15T17:45:44Z) - PREDILECT: Preferences Delineated with Zero-Shot Language-based
Reasoning in Reinforcement Learning [2.7387720378113554]
Preference-based reinforcement learning (RL) has emerged as a new field in robot learning.
We use the zero-shot capabilities of a large language model (LLM) to reason from the text provided by humans.
In both a simulated scenario and a user study, we reveal the effectiveness of our work by analyzing the feedback and its implications.
arXiv Detail & Related papers (2024-02-23T16:30:05Z) - ImitationNet: Unsupervised Human-to-Robot Motion Retargeting via Shared Latent Space [9.806227900768926]
This paper introduces a novel deep-learning approach for human-to-robot motion.
Our method does not require paired human-to-robot data, which facilitates its translation to new robots.
Our model outperforms existing works regarding human-to-robot similarity in terms of efficiency and precision.
arXiv Detail & Related papers (2023-09-11T08:55:04Z) - Learning Video-Conditioned Policies for Unseen Manipulation Tasks [83.2240629060453]
Video-conditioned Policy learning maps human demonstrations of previously unseen tasks to robot manipulation skills.
We learn our policy to generate appropriate actions given current scene observations and a video of the target task.
We validate our approach on a set of challenging multi-task robot manipulation environments and outperform state of the art.
arXiv Detail & Related papers (2023-05-10T16:25:42Z) - Open-World Object Manipulation using Pre-trained Vision-Language Models [72.87306011500084]
For robots to follow instructions from people, they must be able to connect the rich semantic information in human vocabulary.
We develop a simple approach, which leverages a pre-trained vision-language model to extract object-identifying information.
In a variety of experiments on a real mobile manipulator, we find that MOO generalizes zero-shot to a wide range of novel object categories and environments.
arXiv Detail & Related papers (2023-03-02T01:55:10Z) - Zero-Shot Robot Manipulation from Passive Human Videos [59.193076151832145]
We develop a framework for extracting agent-agnostic action representations from human videos.
Our framework is based on predicting plausible human hand trajectories.
We deploy the trained model zero-shot for physical robot manipulation tasks.
arXiv Detail & Related papers (2023-02-03T21:39:52Z) - Signs of Language: Embodied Sign Language Fingerspelling Acquisition
from Demonstrations for Human-Robot Interaction [1.0166477175169308]
We propose an approach for learning dexterous motor imitation from video examples without additional information.
We first build a URDF model of a robotic hand with a single actuator for each joint.
We then leverage pre-trained deep vision models to extract the 3D pose of the hand from RGB videos.
arXiv Detail & Related papers (2022-09-12T10:42:26Z) - Human-to-Robot Imitation in the Wild [50.49660984318492]
We propose an efficient one-shot robot learning algorithm, centered around learning from a third-person perspective.
We show one-shot generalization and success in real-world settings, including 20 different manipulation tasks in the wild.
arXiv Detail & Related papers (2022-07-19T17:59:59Z) - Joint Mind Modeling for Explanation Generation in Complex Human-Robot
Collaborative Tasks [83.37025218216888]
We propose a novel explainable AI (XAI) framework for achieving human-like communication in human-robot collaborations.
The robot builds a hierarchical mind model of the human user and generates explanations of its own mind as a form of communications.
Results show that the generated explanations of our approach significantly improves the collaboration performance and user perception of the robot.
arXiv Detail & Related papers (2020-07-24T23:35:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.