Visibility Guided NMS: Efficient Boosting of Amodal Object Detection in
Crowded Traffic Scenes
- URL: http://arxiv.org/abs/2006.08547v1
- Date: Mon, 15 Jun 2020 17:03:23 GMT
- Title: Visibility Guided NMS: Efficient Boosting of Amodal Object Detection in
Crowded Traffic Scenes
- Authors: Nils G\"ahlert, Niklas Hanselmann, Uwe Franke, Joachim Denzler
- Abstract summary: Modern 2D object detection frameworks predict multiple bounding boxes per object that are refined using Non-Maximum-Suppression (NMS) to suppress all but one bounding box.
Our novel Visibility Guided NMS (vg-NMS) leverages both pixel-based as well as amodal object detection paradigms and improves the detection performance especially for highly occluded objects with little computational overhead.
We evaluate vg-NMS using KITTI, VIPER as well as the Synscapes dataset and show that it outperforms current state-of-the-art NMS.
- Score: 7.998326245039892
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Object detection is an important task in environment perception for
autonomous driving. Modern 2D object detection frameworks such as Yolo, SSD or
Faster R-CNN predict multiple bounding boxes per object that are refined using
Non-Maximum-Suppression (NMS) to suppress all but one bounding box. While
object detection itself is fully end-to-end learnable and does not require any
manual parameter selection, standard NMS is parametrized by an overlap
threshold that has to be chosen by hand. In practice, this often leads to an
inability of standard NMS strategies to distinguish different objects in
crowded scenes in the presence of high mutual occlusion, e.g. for parked cars
or crowds of pedestrians. Our novel Visibility Guided NMS (vg-NMS) leverages
both pixel-based as well as amodal object detection paradigms and improves the
detection performance especially for highly occluded objects with little
computational overhead. We evaluate vg-NMS using KITTI, VIPER as well as the
Synscapes dataset and show that it outperforms current state-of-the-art NMS.
Related papers
- Fuzzy-NMS: Improving 3D Object Detection with Fuzzy Classification in
NMS [19.452760776980472]
Non-maximum suppression (NMS) is an essential post-processing module used in many 3D object detection frameworks.
We introduce fuzzy learning into NMS and propose a novel generalized Fuzzy-NMS module to achieve finer candidate bounding box filtering.
arXiv Detail & Related papers (2023-10-21T09:09:03Z) - DARTH: Holistic Test-time Adaptation for Multiple Object Tracking [87.72019733473562]
Multiple object tracking (MOT) is a fundamental component of perception systems for autonomous driving.
Despite the urge of safety in driving systems, no solution to the MOT adaptation problem to domain shift in test-time conditions has ever been proposed.
We introduce DARTH, a holistic test-time adaptation framework for MOT.
arXiv Detail & Related papers (2023-10-03T10:10:42Z) - RLM-Tracking: Online Multi-Pedestrian Tracking Supported by Relative
Location Mapping [5.9669075749248774]
Problem of multi-object tracking is a fundamental computer vision research focus, widely used in public safety, transport, autonomous vehicles, robotics, and other regions involving artificial intelligence.
In this paper, we design a new multi-object tracker for the above issues that contains an object textbfRelative Location Mapping (RLM) model and textbfTarget Region Density (TRD) model.
The new tracker is more sensitive to the differences in position relationships between objects.
It can introduce low-score detection frames into different regions in real-time according to the density of object
arXiv Detail & Related papers (2022-10-19T11:37:14Z) - CAMO-MOT: Combined Appearance-Motion Optimization for 3D Multi-Object
Tracking with Camera-LiDAR Fusion [34.42289908350286]
3D Multi-object tracking (MOT) ensures consistency during continuous dynamic detection.
It can be challenging to accurately track the irregular motion of objects for LiDAR-based methods.
We propose a novel camera-LiDAR fusion 3D MOT framework based on the Combined Appearance-Motion Optimization (CAMO-MOT)
arXiv Detail & Related papers (2022-09-06T14:41:38Z) - Self-Supervised Object Detection via Generative Image Synthesis [106.65384648377349]
We present the first end-to-end analysis-by synthesis framework with controllable GANs for the task of self-supervised object detection.
We use collections of real world images without bounding box annotations to learn to synthesize and detect objects.
Our work advances the field of self-supervised object detection by introducing a successful new paradigm of using controllable GAN-based image synthesis for it.
arXiv Detail & Related papers (2021-10-19T11:04:05Z) - GrooMeD-NMS: Grouped Mathematically Differentiable NMS for Monocular 3D
Object Detection [25.313894069303718]
We present and integrate GrooMeD-NMS -- a novel Grouped Mathematically Differentiable NMS for monocular 3D object detection.
GrooMeD-NMS addresses the mismatch between training and inference pipelines.
It achieves state-of-the-art monocular 3D object detection results on the KITTI benchmark dataset.
arXiv Detail & Related papers (2021-03-31T16:29:50Z) - Object Detection Made Simpler by Eliminating Heuristic NMS [70.93004137521946]
We show a simple NMS-free, end-to-end object detection framework.
We attain on par or even improved detection accuracy compared with the original one-stage detector.
arXiv Detail & Related papers (2021-01-28T02:38:29Z) - End-to-End Object Detection with Fully Convolutional Network [71.56728221604158]
We introduce a Prediction-aware One-To-One (POTO) label assignment for classification to enable end-to-end detection.
A simple 3D Max Filtering (3DMF) is proposed to utilize the multi-scale features and improve the discriminability of convolutions in the local region.
Our end-to-end framework achieves competitive performance against many state-of-the-art detectors with NMS on COCO and CrowdHuman datasets.
arXiv Detail & Related papers (2020-12-07T09:14:55Z) - Simultaneous Detection and Tracking with Motion Modelling for Multiple
Object Tracking [94.24393546459424]
We introduce Deep Motion Modeling Network (DMM-Net) that can estimate multiple objects' motion parameters to perform joint detection and association.
DMM-Net achieves PR-MOTA score of 12.80 @ 120+ fps for the popular UA-DETRAC challenge, which is better performance and orders of magnitude faster.
We also contribute a synthetic large-scale public dataset Omni-MOT for vehicle tracking that provides precise ground-truth annotations.
arXiv Detail & Related papers (2020-08-20T08:05:33Z) - SoDA: Multi-Object Tracking with Soft Data Association [75.39833486073597]
Multi-object tracking (MOT) is a prerequisite for a safe deployment of self-driving cars.
We propose a novel approach to MOT that uses attention to compute track embeddings that encode dependencies between observed objects.
arXiv Detail & Related papers (2020-08-18T03:40:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.