Full-Body Articulated Human-Object Interaction
- URL: http://arxiv.org/abs/2212.10621v3
- Date: Mon, 18 Dec 2023 15:33:51 GMT
- Title: Full-Body Articulated Human-Object Interaction
- Authors: Nan Jiang, Tengyu Liu, Zhexuan Cao, Jieming Cui, Zhiyuan zhang, Yixin
Chen, He Wang, Yixin Zhu, Siyuan Huang
- Abstract summary: CHAIRS is a large-scale motion-captured f-AHOI dataset consisting of 16.2 hours of versatile interactions.
CHAIRS provides 3D meshes of both humans and articulated objects during the entire interactive process.
By learning the geometrical relationships in HOI, we devise the very first model that leverage human pose estimation.
- Score: 61.01135739641217
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Fine-grained capturing of 3D HOI boosts human activity understanding and
facilitates downstream visual tasks, including action recognition, holistic
scene reconstruction, and human motion synthesis. Despite its significance,
existing works mostly assume that humans interact with rigid objects using only
a few body parts, limiting their scope. In this paper, we address the
challenging problem of f-AHOI, wherein the whole human bodies interact with
articulated objects, whose parts are connected by movable joints. We present
CHAIRS, a large-scale motion-captured f-AHOI dataset, consisting of 16.2 hours
of versatile interactions between 46 participants and 81 articulated and rigid
sittable objects. CHAIRS provides 3D meshes of both humans and articulated
objects during the entire interactive process, as well as realistic and
physically plausible full-body interactions. We show the value of CHAIRS with
object pose estimation. By learning the geometrical relationships in HOI, we
devise the very first model that leverage human pose estimation to tackle the
estimation of articulated object poses and shapes during whole-body
interactions. Given an image and an estimated human pose, our model first
reconstructs the pose and shape of the object, then optimizes the
reconstruction according to a learned interaction prior. Under both evaluation
settings (e.g., with or without the knowledge of objects'
geometries/structures), our model significantly outperforms baselines. We hope
CHAIRS will promote the community towards finer-grained interaction
understanding. We will make the data/code publicly available.
Related papers
- ParaHome: Parameterizing Everyday Home Activities Towards 3D Generative
Modeling of Human-Object Interactions [11.32229757116179]
We introduce the ParaHome system, designed to capture dynamic 3D movements of humans and objects within a common home environment.
By leveraging the ParaHome system, we collect a novel large-scale dataset of human-object interaction.
arXiv Detail & Related papers (2024-01-18T18:59:58Z) - RHOBIN Challenge: Reconstruction of Human Object Interaction [83.07185402102253]
First RHOBIN challenge: reconstruction of human-object interactions in conjunction with the RHOBIN workshop.
Our challenge consists of three tracks of 3D reconstruction from monocular RGB images with a focus on dealing with challenging interaction scenarios.
This paper describes the settings of our challenge and discusses the winning methods of each track in more detail.
arXiv Detail & Related papers (2024-01-07T23:37:07Z) - LEMON: Learning 3D Human-Object Interaction Relation from 2D Images [56.6123961391372]
Learning 3D human-object interaction relation is pivotal to embodied AI and interaction modeling.
Most existing methods approach the goal by learning to predict isolated interaction elements.
We present LEMON, a unified model that mines interaction intentions of the counterparts and employs curvatures to guide the extraction of geometric correlations.
arXiv Detail & Related papers (2023-12-14T14:10:57Z) - CG-HOI: Contact-Guided 3D Human-Object Interaction Generation [29.3564427724612]
We propose CG-HOI, the first method to generate dynamic 3D human-object interactions (HOIs) from text.
We model the motion of both human and object in an interdependent fashion, as semantically rich human motion rarely happens in isolation.
We show that our joint contact-based human-object interaction approach generates realistic and physically plausible sequences.
arXiv Detail & Related papers (2023-11-27T18:59:10Z) - Anthropomorphic Grasping with Neural Object Shape Completion [20.952799332420195]
Humans exhibit extraordinary dexterity when handling objects.
Hand postures commonly demonstrate the influence of specific regions on objects that need to be grasped.
In this work, we leverage human-like object understanding by reconstructing and completing their full geometry from partial observations.
arXiv Detail & Related papers (2023-11-04T21:05:26Z) - Reconstructing Action-Conditioned Human-Object Interactions Using
Commonsense Knowledge Priors [42.17542596399014]
We present a method for inferring diverse 3D models of human-object interactions from images.
Our method extracts high-level commonsense knowledge from large language models.
We quantitatively evaluate the inferred 3D models on a large human-object interaction dataset.
arXiv Detail & Related papers (2022-09-06T13:32:55Z) - Learn to Predict How Humans Manipulate Large-sized Objects from
Interactive Motions [82.90906153293585]
We propose a graph neural network, HO-GCN, to fuse motion data and dynamic descriptors for the prediction task.
We show the proposed network that consumes dynamic descriptors can achieve state-of-the-art prediction results and help the network better generalize to unseen objects.
arXiv Detail & Related papers (2022-06-25T09:55:39Z) - THORN: Temporal Human-Object Relation Network for Action Recognition [3.6704226968275258]
Most action recognition models treat human activities as unitary events.
In this paper we propose to recognize human action by leveraging the set of interactions that define an action.
We present an end-to-end network: THORN, that can leverage important human-object and object-object interactions to predict actions.
arXiv Detail & Related papers (2022-04-20T14:00:24Z) - BEHAVE: Dataset and Method for Tracking Human Object Interactions [105.77368488612704]
We present the first full body human- object interaction dataset with multi-view RGBD frames and corresponding 3D SMPL and object fits along with the annotated contacts between them.
We use this data to learn a model that can jointly track humans and objects in natural environments with an easy-to-use portable multi-camera setup.
arXiv Detail & Related papers (2022-04-14T13:21:19Z) - Object Properties Inferring from and Transfer for Human Interaction
Motions [51.896592493436984]
In this paper, we present a fine-grained action recognition method that learns to infer object properties from human interaction motion alone.
We collect a large number of videos and 3D skeletal motions of the performing actors using an inertial motion capture device.
In particular, we learn to identify the interacting object, by estimating its weight, or its fragility or delicacy.
arXiv Detail & Related papers (2020-08-20T14:36:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.