Learning Representations that Enable Generalization in Assistive Tasks
- URL: http://arxiv.org/abs/2212.03175v1
- Date: Mon, 5 Dec 2022 18:59:16 GMT
- Title: Learning Representations that Enable Generalization in Assistive Tasks
- Authors: Jerry Zhi-Yang He, Aditi Raghunathan, Daniel S. Brown, Zackory
Erickson, Anca D. Dragan
- Abstract summary: We focus on enabling generalization in assistive tasks in which the robot is acting to assist a user.
We find that sim2real methods that encode environment (or population) parameters and work well in tasks that robots do in isolation, do not work well in assistance.
- Score: 45.62648124988644
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent work in sim2real has successfully enabled robots to act in physical
environments by training in simulation with a diverse ''population'' of
environments (i.e. domain randomization). In this work, we focus on enabling
generalization in assistive tasks: tasks in which the robot is acting to assist
a user (e.g. helping someone with motor impairments with bathing or with
scratching an itch). Such tasks are particularly interesting relative to prior
sim2real successes because the environment now contains a human who is also
acting. This complicates the problem because the diversity of human users
(instead of merely physical environment parameters) is more difficult to
capture in a population, thus increasing the likelihood of encountering
out-of-distribution (OOD) human policies at test time. We advocate that
generalization to such OOD policies benefits from (1) learning a good latent
representation for human policies that test-time humans can accurately be
mapped to, and (2) making that representation adaptable with test-time
interaction data, instead of relying on it to perfectly capture the space of
human policies based on the simulated population only. We study how to best
learn such a representation by evaluating on purposefully constructed OOD test
policies. We find that sim2real methods that encode environment (or population)
parameters and work well in tasks that robots do in isolation, do not work well
in assistance. In assistance, it seems crucial to train the representation
based on the history of interaction directly, because that is what the robot
will have access to at test time. Further, training these representations to
then predict human actions not only gives them better structure, but also
enables them to be fine-tuned at test-time, when the robot observes the partner
act. https://adaptive-caregiver.github.io.
Related papers
- Robot Interaction Behavior Generation based on Social Motion Forecasting for Human-Robot Interaction [9.806227900768926]
We propose to model social motion forecasting in a shared human-robot representation space.
ECHO operates in the aforementioned shared space to predict the future motions of the agents encountered in social scenarios.
We evaluate our model in multi-person and human-robot motion forecasting tasks and obtain state-of-the-art performance by a large margin.
arXiv Detail & Related papers (2024-02-07T11:37:14Z) - SynH2R: Synthesizing Hand-Object Motions for Learning Human-to-Robot
Handovers [37.49601724575655]
Vision-based human-to-robot handover is an important and challenging task in human-robot interaction.
We introduce a framework that can generate plausible human grasping motions suitable for training the robot.
This allows us to generate synthetic training and testing data with 100x more objects than previous work.
arXiv Detail & Related papers (2023-11-09T18:57:02Z) - Habitat 3.0: A Co-Habitat for Humans, Avatars and Robots [119.55240471433302]
Habitat 3.0 is a simulation platform for studying collaborative human-robot tasks in home environments.
It addresses challenges in modeling complex deformable bodies and diversity in appearance and motion.
Human-in-the-loop infrastructure enables real human interaction with simulated robots via mouse/keyboard or a VR interface.
arXiv Detail & Related papers (2023-10-19T17:29:17Z) - Quantifying Assistive Robustness Via the Natural-Adversarial Frontier [40.125563987538044]
RIGID is a method for training adversarial human policies that trade off between minimizing robot reward and acting human-like.
On an Assistive Gym task, we use RIGID to analyze the performance of standard collaborative Reinforcement Learning.
We also compare the frontier RIGID identifies with the failures identified in expert adversarial interaction, and with naturally-occurring failures during user interaction.
arXiv Detail & Related papers (2023-10-16T17:34:54Z) - Learning Human-to-Robot Handovers from Point Clouds [63.18127198174958]
We propose the first framework to learn control policies for vision-based human-to-robot handovers.
We show significant performance gains over baselines on a simulation benchmark, sim-to-sim transfer and sim-to-real transfer.
arXiv Detail & Related papers (2023-03-30T17:58:36Z) - Self-Improving Robots: End-to-End Autonomous Visuomotor Reinforcement
Learning [54.636562516974884]
In imitation and reinforcement learning, the cost of human supervision limits the amount of data that robots can be trained on.
In this work, we propose MEDAL++, a novel design for self-improving robotic systems.
The robot autonomously practices the task by learning to both do and undo the task, simultaneously inferring the reward function from the demonstrations.
arXiv Detail & Related papers (2023-03-02T18:51:38Z) - Learning Latent Representations to Co-Adapt to Humans [12.71953776723672]
Non-stationary humans are challenging for robot learners.
In this paper we introduce an algorithmic formalism that enables robots to co-adapt alongside dynamic humans.
arXiv Detail & Related papers (2022-12-19T16:19:24Z) - Human-to-Robot Imitation in the Wild [50.49660984318492]
We propose an efficient one-shot robot learning algorithm, centered around learning from a third-person perspective.
We show one-shot generalization and success in real-world settings, including 20 different manipulation tasks in the wild.
arXiv Detail & Related papers (2022-07-19T17:59:59Z) - Reactive Long Horizon Task Execution via Visual Skill and Precondition
Models [59.76233967614774]
We describe an approach for sim-to-real training that can accomplish unseen robotic tasks using models learned in simulation to ground components of a simple task planner.
We show an increase in success rate from 91.6% to 98% in simulation and from 10% to 80% success rate in the real-world as compared with naive baselines.
arXiv Detail & Related papers (2020-11-17T15:24:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.