YOLOv3 with Spatial Pyramid Pooling for Object Detection with Unmanned
Aerial Vehicles
- URL: http://arxiv.org/abs/2305.12344v1
- Date: Sun, 21 May 2023 04:41:52 GMT
- Title: YOLOv3 with Spatial Pyramid Pooling for Object Detection with Unmanned
Aerial Vehicles
- Authors: Wahyu Pebrianto, Panca Mudjirahardjo, Sholeh Hadi Pramono, Rahmadwati,
Raden Arief Setyawan
- Abstract summary: We aim to improve the performance of the one-stage detector YOLOv3 by adding a Spatial Pyramid Pooling layer on the end of the backbone darknet-53.
We also conducted an evaluation study on different versions of YOLOv3 methods.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Object detection with Unmanned Aerial Vehicles (UAVs) has attracted much
attention in the research field of computer vision. However, not easy to
accurately detect objects with data obtained from UAVs, which capture images
from very high altitudes, making the image dominated by small object sizes,
that difficult to detect. Motivated by that challenge, we aim to improve the
performance of the one-stage detector YOLOv3 by adding a Spatial Pyramid
Pooling (SPP) layer on the end of the backbone darknet-53 to obtain more
efficient feature extraction process in object detection tasks with UAVs. We
also conducted an evaluation study on different versions of YOLOv3 methods.
Includes YOLOv3 with SPP, YOLOv3, and YOLOv3-tiny, which we analyzed with the
VisDrone2019-Det dataset. Here we show that YOLOv3 with SPP can get results mAP
0.6% higher than YOLOv3 and 26.6% than YOLOv3-Tiny at 640x640 input scale and
is even able to maintain accuracy at different input image scales than other
versions of the YOLOv3 method. Those results prove that the addition of SPP
layers to YOLOv3 can be an efficient solution for improving the performance of
the object detection method with data obtained from UAVs.
Related papers
- YOLO-World: Real-Time Open-Vocabulary Object Detection [87.08732047660058]
We introduce YOLO-World, an innovative approach that enhances YOLO with open-vocabulary detection capabilities.
Our method excels in detecting a wide range of objects in a zero-shot manner with high efficiency.
YOLO-World achieves 35.4 AP with 52.0 FPS on V100, which outperforms many state-of-the-art methods in terms of both accuracy and speed.
arXiv Detail & Related papers (2024-01-30T18:59:38Z) - Investigating YOLO Models Towards Outdoor Obstacle Detection For
Visually Impaired People [3.4628430044380973]
Seven different YOLO object detection models were implemented.
YOLOv8 was found to be the best model, which reached a precision of $80%$ and a recall of $68.2%$ on a well-known Obstacle dataset.
YOLO-NAS was found to be suboptimal for the obstacle detection task.
arXiv Detail & Related papers (2023-12-10T13:16:22Z) - YOLO-MS: Rethinking Multi-Scale Representation Learning for Real-time
Object Detection [80.11152626362109]
We provide an efficient and performant object detector, termed YOLO-MS.
We train our YOLO-MS on the MS COCO dataset from scratch without relying on any other large-scale datasets.
Our work can also be used as a plug-and-play module for other YOLO models.
arXiv Detail & Related papers (2023-08-10T10:12:27Z) - YOLO-Drone:Airborne real-time detection of dense small objects from
high-altitude perspective [8.864582442699023]
A real-time object detection algorithm (YOLO-Drone) is proposed and applied to two new UAV platforms and a specific light source.
YOLO-Drone exhibits high real-time inference speed of 53 FPS and a maximum mAP of 34.04%.
Notably, YOLO-Drone achieves high performance under the silicon-based golden LEDs, with a mAP of up to 87.71%.
arXiv Detail & Related papers (2023-04-14T05:21:47Z) - An advanced YOLOv3 method for small object detection [2.906551456030129]
This paper introduces an improved YOLOv3 algorithm for small object detection.
In the proposed method, the dilated convolutions mish (DCM) module is introduced into the backbone network of YOLOv3.
In the neck network of YOLOv3, the convolutional block attention module (CBAM) and multi-level fusion module are introduced.
arXiv Detail & Related papers (2022-12-06T07:58:21Z) - VPIT: Real-time Embedded Single Object 3D Tracking Using Voxel Pseudo Images [90.60881721134656]
We propose a novel voxel-based 3D single object tracking (3D SOT) method called Voxel Pseudo Image Tracking (VPIT)
Experiments on KITTI Tracking dataset show that VPIT is the fastest 3D SOT method and maintains competitive Success and Precision values.
arXiv Detail & Related papers (2022-06-06T14:02:06Z) - SALISA: Saliency-based Input Sampling for Efficient Video Object
Detection [58.22508131162269]
We propose SALISA, a novel non-uniform SALiency-based Input SAmpling technique for video object detection.
We show that SALISA significantly improves the detection of small objects.
arXiv Detail & Related papers (2022-04-05T17:59:51Z) - A lightweight and accurate YOLO-like network for small target detection
in Aerial Imagery [94.78943497436492]
We present YOLO-S, a simple, fast and efficient network for small target detection.
YOLO-S exploits a small feature extractor based on Darknet20, as well as skip connection, via both bypass and concatenation.
YOLO-S has an 87% decrease of parameter size and almost one half FLOPs of YOLOv3, making practical the deployment for low-power industrial applications.
arXiv Detail & Related papers (2022-04-05T16:29:49Z) - Expandable YOLO: 3D Object Detection from RGB-D Images [64.14512458954344]
This paper aims at constructing a light-weight object detector that inputs a depth and a color image from a stereo camera.
By extending the network architecture of YOLOv3 to 3D in the middle, it is possible to output in the depth direction.
Intersection over Uninon (IoU) in 3D space is introduced to confirm the accuracy of region extraction results.
arXiv Detail & Related papers (2020-06-26T07:32:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.