ContactGen: Contact-Guided Interactive 3D Human Generation for Partners
- URL: http://arxiv.org/abs/2401.17212v2
- Date: Sat, 3 Feb 2024 06:22:40 GMT
- Title: ContactGen: Contact-Guided Interactive 3D Human Generation for Partners
- Authors: Dongjun Gu, Jaehyeok Shim, Jaehoon Jang, Changwoo Kang, Kyungdon Joo
- Abstract summary: We introduce a new task of 3D human generation in terms of physical contact.
A given partner human can have diverse poses and different contact regions according to the type of interaction.
We propose a novel method of generating interactive 3D humans for a given partner human based on a guided diffusion framework.
- Score: 9.13466172688693
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Among various interactions between humans, such as eye contact and gestures,
physical interactions by contact can act as an essential moment in
understanding human behaviors. Inspired by this fact, given a 3D partner human
with the desired interaction label, we introduce a new task of 3D human
generation in terms of physical contact. Unlike previous works of interacting
with static objects or scenes, a given partner human can have diverse poses and
different contact regions according to the type of interaction. To handle this
challenge, we propose a novel method of generating interactive 3D humans for a
given partner human based on a guided diffusion framework. Specifically, we
newly present a contact prediction module that adaptively estimates potential
contact regions between two input humans according to the interaction label.
Using the estimated potential contact regions as complementary guidances, we
dynamically enforce ContactGen to generate interactive 3D humans for a given
partner human within a guided diffusion model. We demonstrate ContactGen on the
CHI3D dataset, where our method generates physically plausible and diverse
poses compared to comparison methods.
Related papers
- Diverse 3D Human Pose Generation in Scenes based on Decoupled Structure [2.9895817635228017]
We present a novel method for generating diverse 3D human poses in scenes with semantic control.
Our approach consists of three stages: pose generation, contact generation, and putting human into the scene.
The experimental results on the PROX dataset demonstrate that our method produces more physically plausible interactions.
arXiv Detail & Related papers (2024-06-09T08:33:10Z) - Joint Reconstruction of 3D Human and Object via Contact-Based Refinement Transformer [58.98785899556135]
We present a novel joint 3D human-object reconstruction method (CONTHO) that effectively exploits contact information between humans and objects.
There are two core designs in our system: 1) 3D-guided contact estimation and 2) contact-based 3D human and object refinement.
arXiv Detail & Related papers (2024-04-07T06:01:49Z) - Contact-aware Human Motion Generation from Textual Descriptions [57.871692507044344]
This paper addresses the problem of generating 3D interactive human motion from text.
We create a novel dataset named RICH-CAT, representing "Contact-Aware Texts"
We propose a novel approach named CATMO for text-driven interactive human motion synthesis.
arXiv Detail & Related papers (2024-03-23T04:08:39Z) - Beyond the Contact: Discovering Comprehensive Affordance for 3D Objects from Pre-trained 2D Diffusion Models [8.933560282929726]
We introduce a novel affordance representation, named Comprehensive Affordance (ComA)
Given a 3D object mesh, ComA models the distribution of relative orientation and proximity of vertices in interacting human meshes.
We demonstrate that ComA outperforms competitors that rely on human annotations in modeling contact-based affordance.
arXiv Detail & Related papers (2024-01-23T18:59:59Z) - CG-HOI: Contact-Guided 3D Human-Object Interaction Generation [29.3564427724612]
We propose CG-HOI, the first method to generate dynamic 3D human-object interactions (HOIs) from text.
We model the motion of both human and object in an interdependent fashion, as semantically rich human motion rarely happens in isolation.
We show that our joint contact-based human-object interaction approach generates realistic and physically plausible sequences.
arXiv Detail & Related papers (2023-11-27T18:59:10Z) - InterControl: Zero-shot Human Interaction Generation by Controlling Every Joint [67.6297384588837]
We introduce a novel controllable motion generation method, InterControl, to encourage the synthesized motions maintaining the desired distance between joint pairs.
We demonstrate that the distance between joint pairs for human-wise interactions can be generated using an off-the-shelf Large Language Model.
arXiv Detail & Related papers (2023-11-27T14:32:33Z) - ContactGen: Generative Contact Modeling for Grasp Generation [37.56729700157981]
This paper presents a novel object-centric contact representation ContactGen for hand-object interaction.
We propose a conditional generative model to predict ContactGen and adopt model-based optimization to predict diverse and geometrically feasible grasps.
arXiv Detail & Related papers (2023-10-05T17:59:45Z) - Contact-aware Human Motion Forecasting [87.04827994793823]
We tackle the task of scene-aware 3D human motion forecasting, which consists of predicting future human poses given a 3D scene and a past human motion.
Our approach outperforms the state-of-the-art human motion forecasting and human synthesis methods on both synthetic and real datasets.
arXiv Detail & Related papers (2022-10-08T07:53:19Z) - Reconstructing Action-Conditioned Human-Object Interactions Using
Commonsense Knowledge Priors [42.17542596399014]
We present a method for inferring diverse 3D models of human-object interactions from images.
Our method extracts high-level commonsense knowledge from large language models.
We quantitatively evaluate the inferred 3D models on a large human-object interaction dataset.
arXiv Detail & Related papers (2022-09-06T13:32:55Z) - BEHAVE: Dataset and Method for Tracking Human Object Interactions [105.77368488612704]
We present the first full body human- object interaction dataset with multi-view RGBD frames and corresponding 3D SMPL and object fits along with the annotated contacts between them.
We use this data to learn a model that can jointly track humans and objects in natural environments with an easy-to-use portable multi-camera setup.
arXiv Detail & Related papers (2022-04-14T13:21:19Z) - PLACE: Proximity Learning of Articulation and Contact in 3D Environments [70.50782687884839]
We propose a novel interaction generation method, named PLACE, which explicitly models the proximity between the human body and the 3D scene around it.
Our perceptual study shows that PLACE significantly improves the state-of-the-art method, approaching the realism of real human-scene interaction.
arXiv Detail & Related papers (2020-08-12T21:00:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.