I^2R-Net: Intra- and Inter-Human Relation Network for Multi-Person Pose
Estimation
- URL: http://arxiv.org/abs/2206.10892v1
- Date: Wed, 22 Jun 2022 07:44:41 GMT
- Title: I^2R-Net: Intra- and Inter-Human Relation Network for Multi-Person Pose
Estimation
- Authors: Yiwei Ding, Wenjin Deng, Yinglin Zheng, Pengfei Liu, Meihong Wang,
Xuan Cheng, Jianmin Bao, Dong Chen, Ming Zeng
- Abstract summary: We present the Intra- and Inter-Human Relation Networks (I2R-Net) for Multi-Person Pose Estimation.
First, the Intra-Human Relation Module operates on a single person and aims to capture Intra-Human dependencies.
Second, the Inter-Human Relation Module considers the relation between multiple instances and focuses on capturing Inter-Human interactions.
- Score: 30.204633647947293
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we present the Intra- and Inter-Human Relation Networks
(I^2R-Net) for Multi-Person Pose Estimation. It involves two basic modules.
First, the Intra-Human Relation Module operates on a single person and aims to
capture Intra-Human dependencies. Second, the Inter-Human Relation Module
considers the relation between multiple instances and focuses on capturing
Inter-Human interactions. The Inter-Human Relation Module can be designed very
lightweight by reducing the resolution of feature map, yet learn useful
relation information to significantly boost the performance of the Intra-Human
Relation Module. Even without bells and whistles, our method can compete or
outperform current competition winners. We conduct extensive experiments on
COCO, CrowdPose, and OCHuman datasets. The results demonstrate that the
proposed model surpasses all the state-of-the-art methods. Concretely, the
proposed method achieves 77.4% AP on CrowPose dataset and 67.8% AP on OCHuman
dataset respectively, outperforming existing methods by a large margin.
Additionally, the ablation study and visualization analysis also prove the
effectiveness of our model.
Related papers
- Relation Learning and Aggregate-attention for Multi-person Motion Prediction [13.052342503276936]
Multi-person motion prediction considers not just the skeleton structures or human trajectories but also the interactions between others.
Previous methods often overlook that the joints relations within an individual (intra-relation) and interactions among groups (inter-relation) are distinct types of representations.
We introduce a new collaborative framework for multi-person motion prediction that explicitly modeling these relations.
arXiv Detail & Related papers (2024-11-06T07:48:30Z) - A Plug-and-Play Method for Rare Human-Object Interactions Detection by Bridging Domain Gap [50.079224604394]
We present a novel model-agnostic framework called textbfContext-textbfEnhanced textbfFeature textbfAment (CEFA)
CEFA consists of a feature alignment module and a context enhancement module.
Our method can serve as a plug-and-play module to improve the detection performance of HOI models on rare categories.
arXiv Detail & Related papers (2024-07-31T08:42:48Z) - Closely Interactive Human Reconstruction with Proxemics and Physics-Guided Adaption [64.07607726562841]
Existing multi-person human reconstruction approaches mainly focus on recovering accurate poses or avoiding penetration.
In this work, we tackle the task of reconstructing closely interactive humans from a monocular video.
We propose to leverage knowledge from proxemic behavior and physics to compensate the lack of visual information.
arXiv Detail & Related papers (2024-04-17T11:55:45Z) - Learning Mutual Excitation for Hand-to-Hand and Human-to-Human
Interaction Recognition [22.538114033191313]
We propose a mutual excitation graph convolutional network (me-GCN) by stacking mutual excitation graph convolution layers.
Me-GC learns mutual information in each layer and each stage of graph convolution operations.
Our proposed me-GC outperforms state-of-the-art GCN-based and Transformer-based methods.
arXiv Detail & Related papers (2024-02-04T10:00:00Z) - I2SRM: Intra- and Inter-Sample Relationship Modeling for Multimodal
Information Extraction [10.684005956288347]
We present the Intra- and Inter-Sample Relationship Modeling (I2SRM) method for this task.
Our proposed method achieves competitive results, 77.12% F1-score on Twitter-2015, 88.40% F1-score on Twitter-2017, and 84.12% F1-score on MNRE.
arXiv Detail & Related papers (2023-10-10T05:50:25Z) - Inverse Reinforcement Learning for Text Summarization [52.765898203824975]
We introduce inverse reinforcement learning (IRL) as an effective paradigm for training abstractive summarization models.
Experimental results across datasets in different domains demonstrate the superiority of our proposed IRL model for summarization over MLE and RL baselines.
arXiv Detail & Related papers (2022-12-19T23:45:05Z) - Learning Multiscale Correlations for Human Motion Prediction [10.335804615372629]
We propose a novel multiscale graph convolution network (MGCN) to capture the correlations among human body components.
We evaluate our approach on two standard benchmark datasets for human motion prediction.
arXiv Detail & Related papers (2021-03-19T07:58:16Z) - DecAug: Augmenting HOI Detection via Decomposition [54.65572599920679]
Current algorithms suffer from insufficient training samples and category imbalance within datasets.
We propose an efficient and effective data augmentation method called DecAug for HOI detection.
Experiments show that our method brings up to 3.3 mAP and 1.6 mAP improvements on V-COCO and HICODET dataset.
arXiv Detail & Related papers (2020-10-02T13:59:05Z) - DRG: Dual Relation Graph for Human-Object Interaction Detection [65.50707710054141]
We tackle the challenging problem of human-object interaction (HOI) detection.
Existing methods either recognize the interaction of each human-object pair in isolation or perform joint inference based on complex appearance-based features.
In this paper, we leverage an abstract spatial-semantic representation to describe each human-object pair and aggregate the contextual information of the scene via a dual relation graph.
arXiv Detail & Related papers (2020-08-26T17:59:40Z) - Human Trajectory Forecasting in Crowds: A Deep Learning Perspective [89.4600982169]
We present an in-depth analysis of existing deep learning-based methods for modelling social interactions.
We propose two knowledge-based data-driven methods to effectively capture these social interactions.
We develop a large scale interaction-centric benchmark TrajNet++, a significant yet missing component in the field of human trajectory forecasting.
arXiv Detail & Related papers (2020-07-07T17:19:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.