Wait, That Feels Familiar: Learning to Extrapolate Human Preferences for
Preference Aligned Path Planning
- URL: http://arxiv.org/abs/2309.09912v1
- Date: Mon, 18 Sep 2023 16:24:26 GMT
- Title: Wait, That Feels Familiar: Learning to Extrapolate Human Preferences for
Preference Aligned Path Planning
- Authors: Haresh Karnan, Elvin Yang, Garrett Warnell, Joydeep Biswas, Peter
Stone
- Abstract summary: Preference extrApolation for Terrain awarE Robot Navigation, PATERN, is a novel framework for extrapolating operator terrain preferences for visual navigation.
We show PATERN robustly generalizes to diverse terrains and varied lighting conditions, while navigating in a preference aligned manner.
- Score: 46.66453816117135
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Autonomous mobility tasks such as lastmile delivery require reasoning about
operator indicated preferences over terrains on which the robot should navigate
to ensure both robot safety and mission success. However, coping with out of
distribution data from novel terrains or appearance changes due to lighting
variations remains a fundamental problem in visual terrain adaptive navigation.
Existing solutions either require labor intensive manual data recollection and
labeling or use handcoded reward functions that may not align with operator
preferences. In this work, we posit that operator preferences for visually
novel terrains, which the robot should adhere to, can often be extrapolated
from established terrain references within the inertial, proprioceptive, and
tactile domain. Leveraging this insight, we introduce Preference extrApolation
for Terrain awarE Robot Navigation, PATERN, a novel framework for extrapolating
operator terrain preferences for visual navigation. PATERN learns to map
inertial, proprioceptive, tactile measurements from the robots observations to
a representation space and performs nearest neighbor search in this space to
estimate operator preferences over novel terrains. Through physical robot
experiments in outdoor environments, we assess PATERNs capability to
extrapolate preferences and generalize to novel terrains and challenging
lighting conditions. Compared to baseline approaches, our findings indicate
that PATERN robustly generalizes to diverse terrains and varied lighting
conditions, while navigating in a preference aligned manner.
Related papers
- Watch Your STEPP: Semantic Traversability Estimation using Pose Projected Features [4.392942391043664]
We propose a method for estimating terrain traversability by learning from demonstrations of human walking.
Our approach leverages dense, pixel-wise feature embeddings generated using the DINOv2 vision Transformer model.
By minimizing loss, the network distinguishes between familiar terrain with a low reconstruction error and unfamiliar or hazardous terrain with a higher reconstruction error.
arXiv Detail & Related papers (2025-01-29T11:53:58Z) - Learning autonomous driving from aerial imagery [67.06858775696453]
Photogrammetric simulators allow the synthesis of novel views through the transformation of pre-generated assets into novel views.
We use a Neural Radiance Field (NeRF) as an intermediate representation to synthesize novel views from the point of view of a ground vehicle.
arXiv Detail & Related papers (2024-10-18T05:09:07Z) - JRDB-Traj: A Dataset and Benchmark for Trajectory Forecasting in Crowds [79.00975648564483]
Trajectory forecasting models, employed in fields such as robotics, autonomous vehicles, and navigation, face challenges in real-world scenarios.
This dataset provides comprehensive data, including the locations of all agents, scene images, and point clouds, all from the robot's perspective.
The objective is to predict the future positions of agents relative to the robot using raw sensory input data.
arXiv Detail & Related papers (2023-11-05T18:59:31Z) - Pre-Trained Masked Image Model for Mobile Robot Navigation [16.330708552384053]
2D top-down maps are commonly used for the navigation and exploration of mobile robots through unknown areas.
Recent works have shown that predicting the structural patterns in the environment through learning-based approaches can greatly enhance task efficiency.
We show that the existing foundational vision networks can accomplish the same without any fine-tuning.
arXiv Detail & Related papers (2023-10-10T21:16:29Z) - Learning to Grasp on the Moon from 3D Octree Observations with Deep
Reinforcement Learning [0.0]
This work investigates the applicability of deep reinforcement learning for vision-based robotic grasping of objects on the Moon.
A novel simulation environment with procedurally-generated datasets is created to train agents under challenging conditions.
A model-free off-policy actor-critic algorithm is then employed for end-to-end learning of a policy.
arXiv Detail & Related papers (2022-08-01T12:59:03Z) - Domain and Modality Gaps for LiDAR-based Person Detection on Mobile
Robots [91.01747068273666]
This paper studies existing LiDAR-based person detectors with a particular focus on mobile robot scenarios.
Experiments revolve around the domain gap between driving and mobile robot scenarios, as well as the modality gap between 3D and 2D LiDAR sensors.
Results provide practical insights into LiDAR-based person detection and facilitate informed decisions for relevant mobile robot designs and applications.
arXiv Detail & Related papers (2021-06-21T16:35:49Z) - Domain Adaptation for Outdoor Robot Traversability Estimation from RGB
data with Safety-Preserving Loss [12.697106921197701]
We present an approach based on deep learning to estimate and anticipate the traversing score of different routes in the field of view of an on-board RGB camera.
We then enhance the model's capabilities by addressing domain shifts through gradient-reversal unsupervised adaptation.
Experimental results show that our approach is able to satisfactorily identify traversable areas and to generalize to unseen locations.
arXiv Detail & Related papers (2020-09-16T09:19:33Z) - Visual Navigation Among Humans with Optimal Control as a Supervisor [72.5188978268463]
We propose an approach that combines learning-based perception with model-based optimal control to navigate among humans.
Our approach is enabled by our novel data-generation tool, HumANav.
We demonstrate that the learned navigation policies can anticipate and react to humans without explicitly predicting future human motion.
arXiv Detail & Related papers (2020-03-20T16:13:47Z) - Learning to Move with Affordance Maps [57.198806691838364]
The ability to autonomously explore and navigate a physical space is a fundamental requirement for virtually any mobile autonomous agent.
Traditional SLAM-based approaches for exploration and navigation largely focus on leveraging scene geometry.
We show that learned affordance maps can be used to augment traditional approaches for both exploration and navigation, providing significant improvements in performance.
arXiv Detail & Related papers (2020-01-08T04:05:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.