Gaze-based Attention Recognition for Human-Robot Collaboration
- URL: http://arxiv.org/abs/2303.17619v1
- Date: Thu, 30 Mar 2023 11:55:38 GMT
- Title: Gaze-based Attention Recognition for Human-Robot Collaboration
- Authors: Pooja Prajod, Matteo Lavit Nicora, Matteo Malosio, Elisabeth Andr\'e
- Abstract summary: We present an assembly scenario where a human operator and a cobot collaborate equally to piece together a gearbox.
As a first step, we recognize the areas in the workspace that the human operator is paying attention to.
We propose a novel deep-learning approach to develop an attention recognition model.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Attention (and distraction) recognition is a key factor in improving
human-robot collaboration. We present an assembly scenario where a human
operator and a cobot collaborate equally to piece together a gearbox. The setup
provides multiple opportunities for the cobot to adapt its behavior depending
on the operator's attention, which can improve the collaboration experience and
reduce psychological strain. As a first step, we recognize the areas in the
workspace that the human operator is paying attention to, and consequently,
detect when the operator is distracted. We propose a novel deep-learning
approach to develop an attention recognition model. First, we train a
convolutional neural network to estimate the gaze direction using a publicly
available image dataset. Then, we use transfer learning with a small dataset to
map the gaze direction onto pre-defined areas of interest. Models trained using
this approach performed very well in leave-one-subject-out evaluation on the
small dataset. We performed an additional validation of our models using the
video snippets collected from participants working as an operator in the
presented assembly scenario. Although the recall for the Distracted class was
lower in this case, the models performed well in recognizing the areas the
operator paid attention to. To the best of our knowledge, this is the first
work that validated an attention recognition model using data from a setting
that mimics industrial human-robot collaboration. Our findings highlight the
need for validation of attention recognition solutions in such full-fledged,
non-guided scenarios.
Related papers
- Exploiting Contextual Uncertainty of Visual Data for Efficient Training of Deep Models [0.65268245109828]
We introduce the notion of contextual diversity for active learning CDAL.
We propose a data repair algorithm to curate contextually fair data to reduce model bias.
We are working on developing image retrieval system for wildlife camera trap images and reliable warning system for poor quality rural roads.
arXiv Detail & Related papers (2024-11-04T09:43:33Z) - Modeling User Preferences via Brain-Computer Interfacing [54.3727087164445]
We use Brain-Computer Interfacing technology to infer users' preferences, their attentional correlates towards visual content, and their associations with affective experience.
We link these to relevant applications, such as information retrieval, personalized steering of generative models, and crowdsourcing population estimates of affective experiences.
arXiv Detail & Related papers (2024-05-15T20:41:46Z) - Self-Explainable Affordance Learning with Embodied Caption [63.88435741872204]
We introduce Self-Explainable Affordance learning (SEA) with embodied caption.
SEA enables robots to articulate their intentions and bridge the gap between explainable vision-language caption and visual affordance learning.
We propose a novel model to effectively combine affordance grounding with self-explanation in a simple but efficient manner.
arXiv Detail & Related papers (2024-04-08T15:22:38Z) - EyeTrans: Merging Human and Machine Attention for Neural Code
Summarization [16.694601606682046]
We develop a method for incorporating human attention into machine attention to enhance neural code summarization.
We conduct comprehensive experiments on two code summarization tasks to demonstrate the effectiveness of incorporating human attention into Transformers.
arXiv Detail & Related papers (2024-02-21T19:45:06Z) - Real-time Addressee Estimation: Deployment of a Deep-Learning Model on
the iCub Robot [52.277579221741746]
Addressee Estimation is a skill essential for social robots to interact smoothly with humans.
Inspired by human perceptual skills, a deep-learning model for Addressee Estimation is designed, trained, and deployed on an iCub robot.
The study presents the procedure of such implementation and the performance of the model deployed in real-time human-robot interaction.
arXiv Detail & Related papers (2023-11-09T13:01:21Z) - Enhancing Robot Learning through Learned Human-Attention Feature Maps [6.724036710994883]
We think that embedding auxiliary information about focus point into robot learning would enhance efficiency and robustness of the learning process.
In this paper, we propose a novel approach to model and emulate the human attention with an approximate prediction model.
We test our approach on two learning tasks - object detection and imitation learning.
arXiv Detail & Related papers (2023-08-29T14:23:44Z) - Knowledge Guided Bidirectional Attention Network for Human-Object
Interaction Detection [3.0915392100355192]
We argue that the independent use of the bottom-up parsing strategy in HOI is counter-intuitive and could lead to the diffusion of attention.
We introduce a novel knowledge-guided top-down attention into HOI, and propose to model the relation parsing as a "look and search" process.
We implement the process via unifying the bottom-up and top-down attention in a single encoder-decoder based model.
arXiv Detail & Related papers (2022-07-16T16:42:49Z) - Careful with That! Observation of Human Movements to Estimate Objects
Properties [106.925705883949]
We focus on the features of human motor actions that communicate insights on the weight of an object.
Our final goal is to enable a robot to autonomously infer the degree of care required in object handling.
arXiv Detail & Related papers (2021-03-02T08:14:56Z) - Detecting Human-Object Interactions with Action Co-occurrence Priors [108.31956827512376]
A common problem in human-object interaction (HOI) detection task is that numerous HOI classes have only a small number of labeled examples.
We observe that there exist natural correlations and anti-correlations among human-object interactions.
We present techniques to learn these priors and leverage them for more effective training, especially in rare classes.
arXiv Detail & Related papers (2020-07-17T02:47:45Z) - Attention-Oriented Action Recognition for Real-Time Human-Robot
Interaction [11.285529781751984]
We propose an attention-oriented multi-level network framework to meet the need for real-time interaction.
Specifically, a Pre-Attention network is employed to roughly focus on the interactor in the scene at low resolution.
The other compact CNN receives the extracted skeleton sequence as input for action recognition.
arXiv Detail & Related papers (2020-07-02T12:41:28Z) - Cost-effective Interactive Attention Learning with Neural Attention
Processes [79.8115563067513]
We propose a novel interactive learning framework which we refer to as Interactive Attention Learning (IAL)
IAL is prone to overfitting due to scarcity of human annotations, and requires costly retraining.
We tackle these challenges by proposing a sample-efficient attention mechanism and a cost-effective reranking algorithm for instances and features.
arXiv Detail & Related papers (2020-06-09T17:36:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.