Spatial Concept-Based Navigation with Human Speech Instructions via
Probabilistic Inference on Bayesian Generative Model
- URL: http://arxiv.org/abs/2002.07381v2
- Date: Wed, 26 Aug 2020 09:23:29 GMT
- Title: Spatial Concept-Based Navigation with Human Speech Instructions via
Probabilistic Inference on Bayesian Generative Model
- Authors: Akira Taniguchi, Yoshinobu Hagiwara, Tadahiro Taniguchi, Tetsunari
Inamura
- Abstract summary: The aim of this study is to enable a mobile robot to perform navigational tasks with human speech instructions.
Path planning was formalized as the spatial probabilistic distribution on the path-trajectory under speech instruction.
We demonstrated path planning based on human instruction using acquired spatial concepts to verify the usefulness of the proposed approach in the simulator and in real environments.
- Score: 8.851071399120542
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Robots are required to not only learn spatial concepts autonomously but also
utilize such knowledge for various tasks in a domestic environment. Spatial
concept represents a multimodal place category acquired from the robot's
spatial experience including vision, speech-language, and self-position. The
aim of this study is to enable a mobile robot to perform navigational tasks
with human speech instructions, such as `Go to the kitchen', via probabilistic
inference on a Bayesian generative model using spatial concepts. Specifically,
path planning was formalized as the maximization of probabilistic distribution
on the path-trajectory under speech instruction, based on a
control-as-inference framework. Furthermore, we described the relationship
between probabilistic inference based on the Bayesian generative model and
control problem including reinforcement learning. We demonstrated path planning
based on human instruction using acquired spatial concepts to verify the
usefulness of the proposed approach in the simulator and in real environments.
Experimentally, places instructed by the user's speech commands showed high
probability values, and the trajectory toward the target place was correctly
estimated. Our approach, based on probabilistic inference concerning
decision-making, can lead to further improvement in robot autonomy.
Related papers
- Gaussian Mixture Models for Affordance Learning using Bayesian Networks [50.18477618198277]
Affordances are fundamental descriptors of relationships between actions, objects and effects.
This paper approaches the problem of an embodied agent exploring the world and learning these affordances autonomously from its sensory experiences.
arXiv Detail & Related papers (2024-02-08T22:05:45Z) - Multi-Agent Dynamic Relational Reasoning for Social Robot Navigation [55.65482030032804]
Social robot navigation can be helpful in various contexts of daily life but requires safe human-robot interactions and efficient trajectory planning.
We propose a systematic relational reasoning approach with explicit inference of the underlying dynamically evolving relational structures.
Our approach infers dynamically evolving relation graphs and hypergraphs to capture the evolution of relations, which the trajectory predictor employs to generate future states.
arXiv Detail & Related papers (2024-01-22T18:58:22Z) - Navigation with Large Language Models: Semantic Guesswork as a Heuristic
for Planning [73.0990339667978]
Navigation in unfamiliar environments presents a major challenge for robots.
We use language models to bias exploration of novel real-world environments.
We evaluate LFG in challenging real-world environments and simulated benchmarks.
arXiv Detail & Related papers (2023-10-16T06:21:06Z) - Visual Affordance Prediction for Guiding Robot Exploration [56.17795036091848]
We develop an approach for learning visual affordances for guiding robot exploration.
We use a Transformer-based model to learn a conditional distribution in the latent embedding space of a VQ-VAE.
We show how the trained affordance model can be used for guiding exploration by acting as a goal-sampling distribution, during visual goal-conditioned policy learning in robotic manipulation.
arXiv Detail & Related papers (2023-05-28T17:53:09Z) - Active Exploration based on Information Gain by Particle Filter for
Efficient Spatial Concept Formation [5.350057408744861]
We propose an active inference method, referred to as spatial concept formation with information gain-based active exploration.
This study interprets the robot's action as a selection of destinations to ask the user, What kind of place is this?' in the context of active inference.
Our experiment demonstrated the effectiveness of the SpCoAE in efficiently determining a destination for learning appropriate spatial concepts in home environments.
arXiv Detail & Related papers (2022-11-20T10:12:17Z) - Interpretable Self-Aware Neural Networks for Robust Trajectory
Prediction [50.79827516897913]
We introduce an interpretable paradigm for trajectory prediction that distributes the uncertainty among semantic concepts.
We validate our approach on real-world autonomous driving data, demonstrating superior performance over state-of-the-art baselines.
arXiv Detail & Related papers (2022-11-16T06:28:20Z) - Hierarchical Path-planning from Speech Instructions with Spatial Concept-based Topometric Semantic Mapping [7.332652485849632]
This study aims to realize a hierarchical spatial representation using a topometric semantic map and path planning with speech instructions, including waypoints.
We conducted experiments in home environments using the Toyota Human Support Robot on the SIGVerse simulator and in a lab-office environment with the real robot, Albert.
Navigation experiments using speech instructions with a waypoint demonstrated a performance improvement of SpCoTMHP over the baseline hierarchical path planning method with path costs.
arXiv Detail & Related papers (2022-03-21T09:15:25Z) - Unsupervised Lexical Acquisition of Relative Spatial Concepts Using
Spoken User Utterances [0.0]
A robot with a flexible spoken dialog system must be able to acquire linguistic representation.
Relative spatial concepts are widely used in our daily lives.
It is not obvious which object is a reference object when a robot learns relative spatial concepts.
arXiv Detail & Related papers (2021-06-16T06:44:27Z) - Language Understanding for Field and Service Robots in a Priori Unknown
Environments [29.16936249846063]
This paper provides a novel learning framework that allows field and service robots to interpret and execute natural language instructions.
We use language as a "sensor" -- inferring spatial, topological, and semantic information implicit in natural language utterances.
We incorporate this distribution in a probabilistic language grounding model and infer a distribution over a symbolic representation of the robot's action space.
arXiv Detail & Related papers (2021-05-21T15:13:05Z) - Congestion-aware Multi-agent Trajectory Prediction for Collision
Avoidance [110.63037190641414]
We propose to learn congestion patterns explicitly and devise a novel "Sense--Learn--Reason--Predict" framework.
By decomposing the learning phases into two stages, a "student" can learn contextual cues from a "teacher" while generating collision-free trajectories.
In experiments, we demonstrate that the proposed model is able to generate collision-free trajectory predictions in a synthetic dataset.
arXiv Detail & Related papers (2021-03-26T02:42:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.