Real-Time target detection in maritime scenarios based on YOLOv3 model
- URL: http://arxiv.org/abs/2003.00800v1
- Date: Mon, 10 Feb 2020 15:25:19 GMT
- Title: Real-Time target detection in maritime scenarios based on YOLOv3 model
- Authors: Alessandro Betti, Benedetto Michelozzi, Andrea Bracci and Andrea
Masini
- Abstract summary: A novel ships dataset is proposed consisting of more than 56k images of marine vessels collected by means of web-scraping.
A YOLOv3 single-stage detector based on Keras API is built on top of this dataset.
- Score: 65.35132992156942
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work a novel ships dataset is proposed consisting of more than 56k
images of marine vessels collected by means of web-scraping and including 12
ship categories. A YOLOv3 single-stage detector based on Keras API is built on
top of this dataset. Current results on four categories (cargo ship, naval
ship, oil ship and tug ship) show Average Precision up to 96% for Intersection
over Union (IoU) of 0.5 and satisfactory detection performances up to IoU of
0.8. A Data Analytics GUI service based on QT framework and Darknet-53 engine
is also implemented in order to simplify the deployment process and analyse
massive amount of images even for people without Data Science expertise.
Related papers
- Real-time Ship Recognition and Georeferencing for the Improvement of Maritime Situational Awareness [0.0]
This thesis presents an investigation into leveraging deep learning and computer vision to advance real-time ship recognition and georeferencing.
A novel dataset, ShipSG, is introduced, containing 3,505 images and 11,625 ship masks with corresponding class and geographic position.
A custom real-time segmentation architecture, ScatYOLOv8+CBAM, is designed for the NVIDIA Jetson AGX Xavier embedded system.
arXiv Detail & Related papers (2024-10-07T11:43:42Z) - Low-power Ship Detection in Satellite Images Using Neuromorphic Hardware [1.4330085996657045]
On-board data processing can identify ships and reduce the amount of data sent to the ground.
Most images captured on board contain only bodies of water or land, with the Airbus Ship Detection dataset showing only 22.1% of images containing ships.
We designed a low-power, two-stage system to optimize performance instead of relying on a single complex model.
arXiv Detail & Related papers (2024-06-17T08:36:12Z) - Deep Transformer Network for Monocular Pose Estimation of Ship-Based UAV [0.23408308015481663]
A Transformer Neural Network model is trained to detect 2D keypoints and estimate the 6D pose of each part.
The method has potential applications for ship-based autonomous UAV landing and navigation.
arXiv Detail & Related papers (2024-06-13T16:01:22Z) - Image and AIS Data Fusion Technique for Maritime Computer Vision
Applications [1.482087972733629]
We develop a technique that fuses Automatic Identification System (AIS) data with vessels detected in images to create datasets.
Our approach associates detected ships to their corresponding AIS messages by estimating distance and azimuth.
This technique is useful for creating datasets for waterway traffic management, encounter detection, and surveillance.
arXiv Detail & Related papers (2023-12-07T20:54:49Z) - Ultra-low Power Deep Learning-based Monocular Relative Localization
Onboard Nano-quadrotors [64.68349896377629]
This work presents a novel autonomous end-to-end system that addresses the monocular relative localization, through deep neural networks (DNNs), of two peer nano-drones.
To cope with the ultra-constrained nano-drone platform, we propose a vertically-integrated framework, including dataset augmentation, quantization, and system optimizations.
Experimental results show that our DNN can precisely localize a 10cm-size target nano-drone by employing only low-resolution monochrome images, up to 2m distance.
arXiv Detail & Related papers (2023-03-03T14:14:08Z) - Navya3DSeg -- Navya 3D Semantic Segmentation Dataset & split generation
for autonomous vehicles [63.20765930558542]
3D semantic data are useful for core perception tasks such as obstacle detection and ego-vehicle localization.
We propose a new dataset, Navya 3D (Navya3DSeg), with a diverse label space corresponding to a large scale production grade operational domain.
It contains 23 labeled sequences and 25 supplementary sequences without labels, designed to explore self-supervised and semi-supervised semantic segmentation benchmarks on point clouds.
arXiv Detail & Related papers (2023-02-16T13:41:19Z) - Pushing the Limits of Asynchronous Graph-based Object Detection with
Event Cameras [62.70541164894224]
We introduce several architecture choices which allow us to scale the depth and complexity of such models while maintaining low computation.
Our method runs 3.7 times faster than a dense graph neural network, taking only 8.4 ms per forward pass.
arXiv Detail & Related papers (2022-11-22T15:14:20Z) - KOLOMVERSE: Korea open large-scale image dataset for object detection in the maritime universe [0.5732204366512352]
We present KOLOMVERSE, an open large-scale image dataset for object detection in the maritime domain by KRISO.
We collected 5,845 hours of video data captured from 21 territorial waters of South Korea.
The dataset has images of 3840$times$2160 pixels and to our knowledge, it is by far the largest publicly available dataset for object detection in the maritime domain.
arXiv Detail & Related papers (2022-06-20T16:45:12Z) - A lightweight and accurate YOLO-like network for small target detection
in Aerial Imagery [94.78943497436492]
We present YOLO-S, a simple, fast and efficient network for small target detection.
YOLO-S exploits a small feature extractor based on Darknet20, as well as skip connection, via both bypass and concatenation.
YOLO-S has an 87% decrease of parameter size and almost one half FLOPs of YOLOv3, making practical the deployment for low-power industrial applications.
arXiv Detail & Related papers (2022-04-05T16:29:49Z) - EAutoDet: Efficient Architecture Search for Object Detection [110.99532343155073]
EAutoDet framework can discover practical backbone and FPN architectures for object detection in 1.4 GPU-days.
We propose a kernel reusing technique by sharing the weights of candidate operations on one edge and consolidating them into one convolution.
In particular, the discovered architectures surpass state-of-the-art object detection NAS methods and achieve 40.1 mAP with 120 FPS and 49.2 mAP with 41.3 FPS on COCO test-dev set.
arXiv Detail & Related papers (2022-03-21T05:56:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.