Improving Visual Perception of a Social Robot for Controlled and
In-the-wild Human-robot Interaction
- URL: http://arxiv.org/abs/2403.01766v2
- Date: Tue, 5 Mar 2024 22:55:23 GMT
- Title: Improving Visual Perception of a Social Robot for Controlled and
In-the-wild Human-robot Interaction
- Authors: Wangjie Zhong, Leimin Tian, Duy Tho Le, Hamid Rezatofighi
- Abstract summary: It is unclear how will the objective interaction performance and subjective user experience be influenced when a social robot adopts a deep-learning based visual perception model.
We employ state-of-the-art human perception and tracking models to improve the visual perception function of the Pepper robot.
- Score: 10.260966795508569
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Social robots often rely on visual perception to understand their users and
the environment. Recent advancements in data-driven approaches for computer
vision have demonstrated great potentials for applying deep-learning models to
enhance a social robot's visual perception. However, the high computational
demands of deep-learning methods, as opposed to the more resource-efficient
shallow-learning models, bring up important questions regarding their effects
on real-world interaction and user experience. It is unclear how will the
objective interaction performance and subjective user experience be influenced
when a social robot adopts a deep-learning based visual perception model. We
employed state-of-the-art human perception and tracking models to improve the
visual perception function of the Pepper robot and conducted a controlled lab
study and an in-the-wild human-robot interaction study to evaluate this novel
perception function for following a specific user with other people present in
the scene.
Related papers
- On the Effect of Robot Errors on Human Teaching Dynamics [1.7249361224827533]
We investigate how the presence and severity of robot errors affect three dimensions of human teaching dynamics.
Results show that people tend to spend more time teaching robots with errors.
Our findings offer valuable insights for designing effective interfaces for interactive learning.
arXiv Detail & Related papers (2024-09-15T19:02:34Z) - A Multi-Modal Explainability Approach for Human-Aware Robots in Multi-Party Conversation [39.87346821309096]
We present an addressee estimation model with improved performance in comparison with the previous SOTA.
We also propose several ways to incorporate explainability and transparency in the aforementioned architecture.
arXiv Detail & Related papers (2024-05-20T13:09:32Z) - Real-time Addressee Estimation: Deployment of a Deep-Learning Model on
the iCub Robot [52.277579221741746]
Addressee Estimation is a skill essential for social robots to interact smoothly with humans.
Inspired by human perceptual skills, a deep-learning model for Addressee Estimation is designed, trained, and deployed on an iCub robot.
The study presents the procedure of such implementation and the performance of the model deployed in real-time human-robot interaction.
arXiv Detail & Related papers (2023-11-09T13:01:21Z) - What Matters to You? Towards Visual Representation Alignment for Robot
Learning [81.30964736676103]
When operating in service of people, robots need to optimize rewards aligned with end-user preferences.
We propose Representation-Aligned Preference-based Learning (RAPL), a method for solving the visual representation alignment problem.
arXiv Detail & Related papers (2023-10-11T23:04:07Z) - Human-oriented Representation Learning for Robotic Manipulation [64.59499047836637]
Humans inherently possess generalizable visual representations that empower them to efficiently explore and interact with the environments in manipulation tasks.
We formalize this idea through the lens of human-oriented multi-task fine-tuning on top of pre-trained visual encoders.
Our Task Fusion Decoder consistently improves the representation of three state-of-the-art visual encoders for downstream manipulation policy-learning.
arXiv Detail & Related papers (2023-10-04T17:59:38Z) - Perception for Humanoid Robots [10.560498559084449]
This review summarizes the recent developments and trends in the field of perception in humanoid robots.
Three main areas of application are identified, namely, internal state estimation, external environment estimation, and human robot interaction.
arXiv Detail & Related papers (2023-09-27T12:32:11Z) - Data-driven emotional body language generation for social robotics [58.88028813371423]
In social robotics, endowing humanoid robots with the ability to generate bodily expressions of affect can improve human-robot interaction and collaboration.
We implement a deep learning data-driven framework that learns from a few hand-designed robotic bodily expressions.
The evaluation study found that the anthropomorphism and animacy of the generated expressions are not perceived differently from the hand-designed ones.
arXiv Detail & Related papers (2022-05-02T09:21:39Z) - Cognitive architecture aided by working-memory for self-supervised
multi-modal humans recognition [54.749127627191655]
The ability to recognize human partners is an important social skill to build personalized and long-term human-robot interactions.
Deep learning networks have achieved state-of-the-art results and demonstrated to be suitable tools to address such a task.
One solution is to make robots learn from their first-hand sensory data with self-supervision.
arXiv Detail & Related papers (2021-03-16T13:50:24Z) - Learning Predictive Models From Observation and Interaction [137.77887825854768]
Learning predictive models from interaction with the world allows an agent, such as a robot, to learn about how the world works.
However, learning a model that captures the dynamics of complex skills represents a major challenge.
We propose a method to augment the training set with observational data of other agents, such as humans.
arXiv Detail & Related papers (2019-12-30T01:10:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.