Self-supervised Social Relation Representation for Human Group Detection
- URL: http://arxiv.org/abs/2203.03843v1
- Date: Tue, 8 Mar 2022 04:26:07 GMT
- Title: Self-supervised Social Relation Representation for Human Group Detection
- Authors: Jiacheng Li, Ruize Han, Haomin Yan, Zekun Qian, Wei Feng, Song Wang
- Abstract summary: We propose a new two-stage multi-head framework for human group detection.
In the first stage, we propose a human behavior simulator head to learn the social relation feature embedding.
In the second stage, based on the social relation embedding, we develop a self-attention inspired network for human group detection.
- Score: 18.38523753680367
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Human group detection, which splits crowd of people into groups, is an
important step for video-based human social activity analysis. The core of
human group detection is the human social relation representation and
division.In this paper, we propose a new two-stage multi-head framework for
human group detection. In the first stage, we propose a human behavior
simulator head to learn the social relation feature embedding, which is
self-supervisely trained by leveraging the socially grounded multi-person
behavior relationship. In the second stage, based on the social relation
embedding, we develop a self-attention inspired network for human group
detection. Remarkable performance on two state-of-the-art large-scale
benchmarks, i.e., PANDA and JRDB-Group, verifies the effectiveness of the
proposed framework. Benefiting from the self-supervised social relation
embedding, our method can provide promising results with very few (labeled)
training data. We will release the source code to the public.
Related papers
- Social Support Detection from Social Media Texts [44.096359084699]
Social support, conveyed through a multitude of interactions and platforms such as social media, plays a pivotal role in fostering a sense of belonging.
This paper introduces Social Support Detection (SSD) as a Natural language processing (NLP) task aimed at identifying supportive interactions.
We conducted experiments on a dataset comprising 10,000 YouTube comments.
arXiv Detail & Related papers (2024-11-04T20:23:03Z) - JRDB-Social: A Multifaceted Robotic Dataset for Understanding of Context and Dynamics of Human Interactions Within Social Groups [8.415759777703125]
JRDB-Social fills gaps in human understanding across diverse indoor and outdoor social contexts.
This dataset aims to enhance our grasp of human social dynamics for robotic applications.
arXiv Detail & Related papers (2024-04-06T00:33:39Z) - SocialBench: Sociality Evaluation of Role-Playing Conversational Agents [85.6641890712617]
Large language models (LLMs) have advanced the development of various AI conversational agents.
SocialBench is the first benchmark designed to evaluate the sociality of role-playing conversational agents at both individual and group levels.
We find that agents excelling in individual level does not imply their proficiency in group level.
arXiv Detail & Related papers (2024-03-20T15:38:36Z) - Multi-Agent Dynamic Relational Reasoning for Social Robot Navigation [50.01551945190676]
Social robot navigation can be helpful in various contexts of daily life but requires safe human-robot interactions and efficient trajectory planning.
We propose a systematic relational reasoning approach with explicit inference of the underlying dynamically evolving relational structures.
We demonstrate its effectiveness for multi-agent trajectory prediction and social robot navigation.
arXiv Detail & Related papers (2024-01-22T18:58:22Z) - Social Motion Prediction with Cognitive Hierarchies [19.71780279070757]
We introduce a new benchmark, a novel formulation, and a cognition-inspired framework.
We present Wusi, a 3D multi-person motion dataset under the context of team sports.
We develop a cognitive hierarchy framework to predict strategic human social interactions.
arXiv Detail & Related papers (2023-11-08T14:51:17Z) - MultiMediate'23: Engagement Estimation and Bodily Behaviour Recognition
in Social Interactions [42.94144353625103]
We address two key human social behaviour analysis tasks for the first time in a controlled challenge: engagement estimation and bodily behaviour recognition in social interactions.
This paper describes the MultiMediate'23 challenge and presents novel sets of annotations for both tasks.
arXiv Detail & Related papers (2023-08-16T09:47:52Z) - Flexible social inference facilitates targeted social learning when
rewards are not observable [58.762004496858836]
Groups coordinate more effectively when individuals are able to learn from others' successes.
We suggest that social inference capacities may help bridge this gap, allowing individuals to update their beliefs about others' underlying knowledge and success from observable trajectories of behavior.
arXiv Detail & Related papers (2022-12-01T21:04:03Z) - Co-Located Human-Human Interaction Analysis using Nonverbal Cues: A
Survey [71.43956423427397]
We aim to identify the nonverbal cues and computational methodologies resulting in effective performance.
This survey differs from its counterparts by involving the widest spectrum of social phenomena and interaction settings.
Some major observations are: the most often used nonverbal cue, computational method, interaction environment, and sensing approach are speaking activity, support vector machines, and meetings composed of 3-4 persons equipped with microphones and cameras, respectively.
arXiv Detail & Related papers (2022-07-20T13:37:57Z) - PHASE: PHysically-grounded Abstract Social Events for Machine Social
Perception [50.551003004553806]
We create a dataset of physically-grounded abstract social events, PHASE, that resemble a wide range of real-life social interactions.
Phase is validated with human experiments demonstrating that humans perceive rich interactions in the social events.
As a baseline model, we introduce a Bayesian inverse planning approach, SIMPLE, which outperforms state-of-the-art feed-forward neural networks.
arXiv Detail & Related papers (2021-03-02T18:44:57Z) - Let me join you! Real-time F-formation recognition by a socially aware
robot [2.8101673772585745]
This paper presents a novel architecture to detect social groups in real-time from a continuous image stream of an ego-vision camera.
We detect F-formations in social gatherings such as meetings, discussions, etc. and predict the robot's approach angle if it wants to join the social group.
We also detect outliers, i.e., the persons who are not part of the group under consideration.
arXiv Detail & Related papers (2020-08-23T17:46:08Z) - Human Trajectory Forecasting in Crowds: A Deep Learning Perspective [89.4600982169]
We present an in-depth analysis of existing deep learning-based methods for modelling social interactions.
We propose two knowledge-based data-driven methods to effectively capture these social interactions.
We develop a large scale interaction-centric benchmark TrajNet++, a significant yet missing component in the field of human trajectory forecasting.
arXiv Detail & Related papers (2020-07-07T17:19:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.