Learning Gaussian Maps for Dense Object Detection
- URL: http://arxiv.org/abs/2004.11855v2
- Date: Thu, 30 Apr 2020 09:51:10 GMT
- Title: Learning Gaussian Maps for Dense Object Detection
- Authors: Sonaal Kant
- Abstract summary: We review common and highly accurate object detection methods on the scenes where numerous similar looking objects are placed in close proximity with each other.
We show that, multi-task learning of gaussian maps along with classification and bounding box regression gives us a significant boost in accuracy over the baseline.
Our method also achieves the state of the art accuracy on the SKU110K citesku110k dataset.
- Score: 1.8275108630751844
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Object detection is a famous branch of research in computer vision, many
state of the art object detection algorithms have been introduced in the recent
past, but how good are those object detectors when it comes to dense object
detection? In this paper we review common and highly accurate object detection
methods on the scenes where numerous similar looking objects are placed in
close proximity with each other. We also show that, multi-task learning of
gaussian maps along with classification and bounding box regression gives us a
significant boost in accuracy over the baseline. We introduce Gaussian Layer
and Gaussian Decoder in the existing RetinaNet network for better accuracy in
dense scenes, with the same computational cost as the RetinaNet. We show the
gain of 6\% and 5\% in mAP with respect to baseline RetinaNet. Our method also
achieves the state of the art accuracy on the SKU110K \cite{sku110k} dataset.
Related papers
- Innovative Horizons in Aerial Imagery: LSKNet Meets DiffusionDet for
Advanced Object Detection [55.2480439325792]
We present an in-depth evaluation of an object detection model that integrates the LSKNet backbone with the DiffusionDet head.
The proposed model achieves a mean average precision (MAP) of approximately 45.7%, which is a significant improvement.
This advancement underscores the effectiveness of the proposed modifications and sets a new benchmark in aerial image analysis.
arXiv Detail & Related papers (2023-11-21T19:49:13Z) - G-CAME: Gaussian-Class Activation Mapping Explainer for Object Detectors [0.0]
G-CAME generates a saliency map as the explanation for object detection models.
We evaluated our method with YOLOX on the MS-COCO 2017 dataset and guided to apply G-CAME into the two-stage Faster-RCNN model.
arXiv Detail & Related papers (2023-06-06T04:30:18Z) - SalienDet: A Saliency-based Feature Enhancement Algorithm for Object
Detection for Autonomous Driving [160.57870373052577]
We propose a saliency-based OD algorithm (SalienDet) to detect unknown objects.
Our SalienDet utilizes a saliency-based algorithm to enhance image features for object proposal generation.
We design a dataset relabeling approach to differentiate the unknown objects from all objects in training sample set to achieve Open-World Detection.
arXiv Detail & Related papers (2023-05-11T16:19:44Z) - Long Range Object-Level Monocular Depth Estimation for UAVs [0.0]
We propose several novel extensions to state-of-the-art methods for monocular object detection from images at long range.
Firstly, we propose Sigmoid and ReLU-like encodings when modeling depth estimation as a regression task.
Secondly, we frame the depth estimation as a classification problem and introduce a Soft-Argmax function in the calculation of the training loss.
arXiv Detail & Related papers (2023-02-17T15:26:04Z) - Robust Object Detection in Remote Sensing Imagery with Noisy and Sparse
Geo-Annotations (Full Version) [4.493174773769076]
In this paper, we present a novel approach for training object detectors with extremely noisy and incomplete annotations.
Our method is based on a teacher-student learning framework and a correction module accounting for imprecise and missing annotations.
We demonstrate that our approach improves standard detectors by 37.1% $AP_50$ on a noisy real-world remote-sensing dataset.
arXiv Detail & Related papers (2022-10-24T07:25:31Z) - Rethinking Spatial Invariance of Convolutional Networks for Object
Counting [119.83017534355842]
We try to use locally connected Gaussian kernels to replace the original convolution filter to estimate the spatial position in the density map.
Inspired by previous work, we propose a low-rank approximation accompanied with translation invariance to favorably implement the approximation of massive Gaussian convolution.
Our methods significantly outperform other state-of-the-art methods and achieve promising learning of the spatial position of objects.
arXiv Detail & Related papers (2022-06-10T17:51:25Z) - You Better Look Twice: a new perspective for designing accurate
detectors with reduced computations [56.34005280792013]
BLT-net is a new low-computation two-stage object detection architecture.
It reduces computations by separating objects from background using a very lite first-stage.
Resulting image proposals are then processed in the second-stage by a highly accurate model.
arXiv Detail & Related papers (2021-07-21T12:39:51Z) - Class agnostic moving target detection by color and location prediction
of moving area [11.326363150470204]
Moving target detection plays an important role in computer vision.
Recent algorithms such as deep learning-based convolutional neural networks have achieved high accuracy and real-time performance.
We propose a model free moving target detection algorithm.
arXiv Detail & Related papers (2021-06-24T12:34:58Z) - A Self-Training Approach for Point-Supervised Object Detection and
Counting in Crowds [54.73161039445703]
We propose a novel self-training approach that enables a typical object detector trained only with point-level annotations.
During training, we utilize the available point annotations to supervise the estimation of the center points of objects.
Experimental results show that our approach significantly outperforms state-of-the-art point-supervised methods under both detection and counting tasks.
arXiv Detail & Related papers (2020-07-25T02:14:42Z) - Rethinking Localization Map: Towards Accurate Object Perception with
Self-Enhancement Maps [78.2581910688094]
This work introduces a novel self-enhancement method to harvest accurate object localization maps and object boundaries with only category labels as supervision.
In particular, the proposed Self-Enhancement Maps achieve the state-of-the-art localization accuracy of 54.88% on ILSVRC.
arXiv Detail & Related papers (2020-06-09T12:35:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.