Localizing Firearm Carriers by Identifying Human-Object Pairs
- URL: http://arxiv.org/abs/2005.09329v2
- Date: Wed, 20 May 2020 09:49:30 GMT
- Title: Localizing Firearm Carriers by Identifying Human-Object Pairs
- Authors: Abdul Basit, Muhammad Akhtar Munir, Mohsen Ali, Arif Mahmood
- Abstract summary: In a given image, human and firearms are separately detected. Each detected human is paired with each detected firearm, allowing us to create a paired bounding box that contains both object and the human.
A network is trained to classify these paired-bounding-boxes into human carrying the identified firearm or not.
- Score: 12.425090880770977
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Visual identification of gunmen in a crowd is a challenging problem, that
requires resolving the association of a person with an object (firearm). We
present a novel approach to address this problem, by defining human-object
interaction (and non-interaction) bounding boxes. In a given image, human and
firearms are separately detected. Each detected human is paired with each
detected firearm, allowing us to create a paired bounding box that contains
both object and the human. A network is trained to classify these
paired-bounding-boxes into human carrying the identified firearm or not.
Extensive experiments were performed to evaluate effectiveness of the
algorithm, including exploiting full pose of the human, hand key-points, and
their association with the firearm. The knowledge of spatially localized
features is key to success of our method by using multi-size proposals with
adaptive average pooling. We have also extended a previously firearm detection
dataset, by adding more images and tagging in extended dataset the
human-firearm pairs (including bounding boxes for firearms and gunmen). The
experimental results ($AP_{hold} = 78.5$) demonstrate effectiveness of the
proposed method.
Related papers
- Detection and Localization of Firearm Carriers in Complex Scenes for
Improved Safety Measures [17.03422574286601]
We propose a novel approach that leverages human-firearm interaction information, which provides valuable clues for localizing firearm carriers.
Our approach incorporates an attention mechanism that effectively distinguishes humans and firearms from the background by focusing on relevant areas.
To handle inputs of varying sizes, we pass paired human-firearm instances with attention masks as channels through a deep network for feature computation.
arXiv Detail & Related papers (2023-09-17T10:50:46Z) - HODN: Disentangling Human-Object Feature for HOI Detection [51.48164941412871]
We propose a Human and Object Disentangling Network (HODN) to model the Human-Object Interaction (HOI) relationships explicitly.
Considering that human features are more contributive to interaction, we propose a Human-Guide Linking method to make sure the interaction decoder focuses on the human-centric regions.
Our proposed method achieves competitive performance on both the V-COCO and the HICO-Det Linking datasets.
arXiv Detail & Related papers (2023-08-20T04:12:50Z) - Among Us: Adversarially Robust Collaborative Perception by Consensus [50.73128191202585]
Multiple robots could perceive a scene (e.g., detect objects) collaboratively better than individuals.
We propose ROBOSAC, a novel sampling-based defense strategy generalizable to unseen attackers.
We validate our method on the task of collaborative 3D object detection in autonomous driving scenarios.
arXiv Detail & Related papers (2023-03-16T17:15:25Z) - Explicit Box Detection Unifies End-to-End Multi-Person Pose Estimation [24.973118696495977]
This paper presents a novel end-to-end framework withExplicit box Detection for multi-person Pose estimation, called ED-Pose.
It unifies the contextual learning between human-level (global) and keypoint-level (local) information.
For the first time, as a fully end-to-end framework with a L1 regression loss, ED-Pose surpasses heatmap-based Top-down methods under the same backbone.
arXiv Detail & Related papers (2023-02-03T08:18:34Z) - HOTR: End-to-End Human-Object Interaction Detection with Transformers [26.664864824357164]
We present a novel framework, referred to by HOTR, which directly predicts a set of human, object, interaction> triplets from an image.
Our proposed algorithm achieves the state-of-the-art performance in two HOI detection benchmarks with an inference time under 1 ms after object detection.
arXiv Detail & Related papers (2021-04-28T10:10:29Z) - Handgun detection using combined human pose and weapon appearance [0.4141979929350861]
In this work, a novel method is proposed to combine, in a single architecture, both weapon appearance and human pose information.
Results obtained show that the combined model improves the handgun detection state of the art, achieving from 4.23 to 18.9 AP points more than the best previous approach.
arXiv Detail & Related papers (2020-10-26T17:45:12Z) - DRG: Dual Relation Graph for Human-Object Interaction Detection [65.50707710054141]
We tackle the challenging problem of human-object interaction (HOI) detection.
Existing methods either recognize the interaction of each human-object pair in isolation or perform joint inference based on complex appearance-based features.
In this paper, we leverage an abstract spatial-semantic representation to describe each human-object pair and aggregate the contextual information of the scene via a dual relation graph.
arXiv Detail & Related papers (2020-08-26T17:59:40Z) - ConsNet: Learning Consistency Graph for Zero-Shot Human-Object
Interaction Detection [101.56529337489417]
We consider the problem of Human-Object Interaction (HOI) Detection, which aims to locate and recognize HOI instances in the form of human, action, object> in images.
We argue that multi-level consistencies among objects, actions and interactions are strong cues for generating semantic representations of rare or previously unseen HOIs.
Our model takes visual features of candidate human-object pairs and word embeddings of HOI labels as inputs, maps them into visual-semantic joint embedding space and obtains detection results by measuring their similarities.
arXiv Detail & Related papers (2020-08-14T09:11:18Z) - Learning Human-Object Interaction Detection using Interaction Points [140.0200950601552]
We propose a novel fully-convolutional approach that directly detects the interactions between human-object pairs.
Our network predicts interaction points, which directly localize and classify the inter-action.
Experiments are performed on two popular benchmarks: V-COCO and HICO-DET.
arXiv Detail & Related papers (2020-03-31T08:42:06Z) - PPDM: Parallel Point Detection and Matching for Real-time Human-Object
Interaction Detection [85.75935399090379]
We propose a single-stage Human-Object Interaction (HOI) detection method that has outperformed all existing methods on HICO-DET dataset at 37 fps.
It is the first real-time HOI detection method.
arXiv Detail & Related papers (2019-12-30T12:00:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.