Do Pedestrians Pay Attention? Eye Contact Detection in the Wild
- URL: http://arxiv.org/abs/2112.04212v1
- Date: Wed, 8 Dec 2021 10:21:28 GMT
- Title: Do Pedestrians Pay Attention? Eye Contact Detection in the Wild
- Authors: Younes Belkada, Lorenzo Bertoni, Romain Caristan, Taylor Mordan and
Alexandre Alahi
- Abstract summary: In urban environments, humans rely on eye contact for fast and efficient communication with nearby people.
In this paper, we focus on eye contact detection in the wild, i.e., real-world scenarios for autonomous vehicles with no control over the environment or the distance of pedestrians.
We introduce a model that leverages semantic keypoints to detect eye contact and show that this high-level representation achieves state-of-the-art results on the publicly-available dataset JAAD.
To study domain adaptation, we create LOOK: a large-scale dataset for eye contact detection in the wild, which focuses on diverse and un
- Score: 75.54077277681353
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In urban or crowded environments, humans rely on eye contact for fast and
efficient communication with nearby people. Autonomous agents also need to
detect eye contact to interact with pedestrians and safely navigate around
them. In this paper, we focus on eye contact detection in the wild, i.e.,
real-world scenarios for autonomous vehicles with no control over the
environment or the distance of pedestrians. We introduce a model that leverages
semantic keypoints to detect eye contact and show that this high-level
representation (i) achieves state-of-the-art results on the publicly-available
dataset JAAD, and (ii) conveys better generalization properties than leveraging
raw images in an end-to-end network. To study domain adaptation, we create
LOOK: a large-scale dataset for eye contact detection in the wild, which
focuses on diverse and unconstrained scenarios for real-world generalization.
The source code and the LOOK dataset are publicly shared towards an open
science mission.
Related papers
- Multi-person eye tracking for real-world scene perception in social settings [34.82692226532414]
We apply mobile eye tracking in a real-world multi-person setup and develop a system to stream, record, and analyse synchronised data.
Our system achieves precise time synchronisation and accurate gaze projection in challenging dynamic scenes.
This advancement enables insight into collaborative behaviour, group dynamics, and social interaction, with high ecological validity.
arXiv Detail & Related papers (2024-07-08T19:33:17Z) - NITEC: Versatile Hand-Annotated Eye Contact Dataset for Ego-Vision
Interaction [2.594420805049218]
We present NITEC, a hand-annotated eye contact dataset for ego-vision interaction.
NITEC exceeds existing datasets for ego-vision eye contact in size and variety of demographics, social contexts, and lighting conditions.
We make our dataset publicly available to foster further exploration in the field of ego-vision interaction.
arXiv Detail & Related papers (2023-11-08T07:42:31Z) - DECO: Dense Estimation of 3D Human-Scene Contact In The Wild [54.44345845842109]
We train a novel 3D contact detector that uses both body-part-driven and scene-context-driven attention to estimate contact on the SMPL body.
We significantly outperform existing SOTA methods across all benchmarks.
We also show qualitatively that DECO generalizes well to diverse and challenging real-world human interactions in natural images.
arXiv Detail & Related papers (2023-09-26T21:21:07Z) - GIMO: Gaze-Informed Human Motion Prediction in Context [75.52839760700833]
We propose a large-scale human motion dataset that delivers high-quality body pose sequences, scene scans, and ego-centric views with eye gaze.
Our data collection is not tied to specific scenes, which further boosts the motion dynamics observed from our subjects.
To realize the full potential of gaze, we propose a novel network architecture that enables bidirectional communication between the gaze and motion branches.
arXiv Detail & Related papers (2022-04-20T13:17:39Z) - GTNet:Guided Transformer Network for Detecting Human-Object Interactions [10.809778265707916]
The human-object interaction (HOI) detection task refers to localizing humans, localizing objects, and predicting the interactions between each human-object pair.
For detecting HOI, it is important to utilize relative spatial configurations and object semantics to find salient spatial regions of images.
This issue is addressed by the novel self-attention based guided transformer network, GTNet.
arXiv Detail & Related papers (2021-08-02T02:06:33Z) - MutualEyeContact: A conversation analysis tool with focus on eye contact [69.17395873398196]
MutualEyeContact can help scientists to understand the importance of (mutual) eye contact in social interactions.
We combine state-of-the-art eye tracking with face recognition based on machine learning and provide a tool for analysis and visualization of social interaction sessions.
arXiv Detail & Related papers (2021-07-09T15:05:53Z) - Onfocus Detection: Identifying Individual-Camera Eye Contact from
Unconstrained Images [81.64699115587167]
Onfocus detection aims at identifying whether the focus of the individual captured by a camera is on the camera or not.
We build a large-scale onfocus detection dataset, named as the OnFocus Detection In the Wild (OFDIW)
We propose a novel end-to-end deep model, i.e., the eye-context interaction inferring network (ECIIN) for onfocus detection.
arXiv Detail & Related papers (2021-03-29T03:29:09Z) - Learning Human-Object Interaction Detection using Interaction Points [140.0200950601552]
We propose a novel fully-convolutional approach that directly detects the interactions between human-object pairs.
Our network predicts interaction points, which directly localize and classify the inter-action.
Experiments are performed on two popular benchmarks: V-COCO and HICO-DET.
arXiv Detail & Related papers (2020-03-31T08:42:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.