Transferable Interactiveness Knowledge for Human-Object Interaction
Detection
- URL: http://arxiv.org/abs/2101.10292v3
- Date: Wed, 3 Mar 2021 10:04:29 GMT
- Title: Transferable Interactiveness Knowledge for Human-Object Interaction
Detection
- Authors: Yong-Lu Li, Xinpeng Liu, Xiaoqian Wu, Xijie Huang, Liang Xu, Cewu Lu
- Abstract summary: We explore interactiveness knowledge which indicates whether a human and an object interact with each other or not.
We found that interactiveness knowledge can be learned across HOI datasets and bridge the gap between diverse HOI category settings.
Our core idea is to exploit an interactiveness network to learn the general interactiveness knowledge from multiple HOI datasets.
- Score: 46.89715038756862
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Human-Object Interaction (HOI) detection is an important problem to
understand how humans interact with objects. In this paper, we explore
interactiveness knowledge which indicates whether a human and an object
interact with each other or not. We found that interactiveness knowledge can be
learned across HOI datasets and bridge the gap between diverse HOI category
settings. Our core idea is to exploit an interactiveness network to learn the
general interactiveness knowledge from multiple HOI datasets and perform
Non-Interaction Suppression (NIS) before HOI classification in inference. On
account of the generalization ability of interactiveness, interactiveness
network is a transferable knowledge learner and can be cooperated with any HOI
detection models to achieve desirable results. We utilize the human instance
and body part features together to learn the interactiveness in hierarchical
paradigm, i.e., instance-level and body part-level interactivenesses.
Thereafter, a consistency task is proposed to guide the learning and extract
deeper interactive visual clues. We extensively evaluate the proposed method on
HICO-DET, V-COCO, and a newly constructed PaStaNet-HOI dataset. With the
learned interactiveness, our method outperforms state-of-the-art HOI detection
methods, verifying its efficacy and flexibility. Code is available at
https://github.com/DirtyHarryLYL/Transferable-Interactiveness-Network.
Related papers
- Visual-Geometric Collaborative Guidance for Affordance Learning [63.038406948791454]
We propose a visual-geometric collaborative guided affordance learning network that incorporates visual and geometric cues.
Our method outperforms the representative models regarding objective metrics and visual quality.
arXiv Detail & Related papers (2024-10-15T07:35:51Z) - Disentangled Interaction Representation for One-Stage Human-Object
Interaction Detection [70.96299509159981]
Human-Object Interaction (HOI) detection is a core task for human-centric image understanding.
Recent one-stage methods adopt a transformer decoder to collect image-wide cues that are useful for interaction prediction.
Traditional two-stage methods benefit significantly from their ability to compose interaction features in a disentangled and explainable manner.
arXiv Detail & Related papers (2023-12-04T08:02:59Z) - HODN: Disentangling Human-Object Feature for HOI Detection [51.48164941412871]
We propose a Human and Object Disentangling Network (HODN) to model the Human-Object Interaction (HOI) relationships explicitly.
Considering that human features are more contributive to interaction, we propose a Human-Guide Linking method to make sure the interaction decoder focuses on the human-centric regions.
Our proposed method achieves competitive performance on both the V-COCO and the HICO-Det Linking datasets.
arXiv Detail & Related papers (2023-08-20T04:12:50Z) - Mining Cross-Person Cues for Body-Part Interactiveness Learning in HOI
Detection [39.61023122191333]
Human-Object Interaction (HOI) detection plays a crucial role in activity understanding.
Previous works only focus on the target person once (i.e., in a local perspective) and overlook the information of the other persons.
In this paper, we argue that comparing body-parts of multi-person simultaneously can afford us more useful and supplementary interactiveness cues.
We construct body-part saliency maps based on self-attention to mine cross-person informative cues and learn the holistic relationships between all the body-parts.
arXiv Detail & Related papers (2022-07-28T15:57:51Z) - Detecting Human-to-Human-or-Object (H2O) Interactions with DIABOLO [29.0200561485714]
We propose a new interaction dataset to deal with both types of human interactions: Human-to-Human-or-Object (H2O)
In addition, we introduce a novel taxonomy of verbs, intended to be closer to a description of human body attitude in relation to the surrounding targets of interaction.
We propose DIABOLO, an efficient subject-centric single-shot method to detect all interactions in one forward pass.
arXiv Detail & Related papers (2022-01-07T11:00:11Z) - Skeleton-Based Mutually Assisted Interacted Object Localization and
Human Action Recognition [111.87412719773889]
We propose a joint learning framework for "interacted object localization" and "human action recognition" based on skeleton data.
Our method achieves the best or competitive performance with the state-of-the-art methods for human action recognition.
arXiv Detail & Related papers (2021-10-28T10:09:34Z) - Learning Human-Object Interaction Detection using Interaction Points [140.0200950601552]
We propose a novel fully-convolutional approach that directly detects the interactions between human-object pairs.
Our network predicts interaction points, which directly localize and classify the inter-action.
Experiments are performed on two popular benchmarks: V-COCO and HICO-DET.
arXiv Detail & Related papers (2020-03-31T08:42:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.