Video object tracking based on YOLOv7 and DeepSORT
- URL: http://arxiv.org/abs/2207.12202v1
- Date: Mon, 25 Jul 2022 13:43:34 GMT
- Title: Video object tracking based on YOLOv7 and DeepSORT
- Authors: Feng Yang, Xingle Zhang, Bo Liu
- Abstract summary: We propose YOLOv7 as the object detection part to the DeepSORT, and propose YOLOv7-DeepSORT.
After experimental evaluation, compared with the previous YOLOv5-DeepSORT, YOLOv7-DeepSORT performances better in tracking accuracy.
- Score: 7.651368436751519
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multiple object tracking (MOT) is an important technology in the field of
computer vision, which is widely used in automatic driving, intelligent
monitoring, behavior recognition and other directions. Among the current
popular MOT methods based on deep learning, Detection Based Tracking (DBT) is
the most widely used in industry, and the performance of them depend on their
object detection network. At present, the DBT algorithm with good performance
and the most widely used is YOLOv5-DeepSORT. Inspired by YOLOv5-DeepSORT, with
the proposal of YOLOv7 network, which performs better in object detection, we
apply YOLOv7 as the object detection part to the DeepSORT, and propose
YOLOv7-DeepSORT. After experimental evaluation, compared with the previous
YOLOv5-DeepSORT, YOLOv7-DeepSORT performances better in tracking accuracy.
Related papers
- Performance Evaluation of Deep Learning-based Quadrotor UAV Detection and Tracking Methods [0.0]
Unmanned Aerial Vehicles (UAVs) are becoming more popular in various sectors, offering many benefits, yet introducing significant challenges to privacy and safety.
This paper investigates state-of-the-art solutions for detecting and tracking quadrotor UAVs to address these concerns.
Cutting-edge deep learning models, specifically the YOLOv5 and YOLOv8 series, are evaluated for their performance in identifying UAVs accurately and quickly.
arXiv Detail & Related papers (2024-09-30T23:37:02Z) - YOLOv5, YOLOv8 and YOLOv10: The Go-To Detectors for Real-time Vision [0.6662800021628277]
This paper focuses on the evolution of the YOLO (You Only Look Once) object detection algorithm, focusing on YOLOv5, YOLOv8, and YOLOv10.
We analyze the architectural advancements, performance improvements, and suitability for edge deployment across these versions.
arXiv Detail & Related papers (2024-07-03T10:40:20Z) - YOLOv10: Real-Time End-to-End Object Detection [68.28699631793967]
YOLOs have emerged as the predominant paradigm in the field of real-time object detection.
The reliance on the non-maximum suppression (NMS) for post-processing hampers the end-to-end deployment of YOLOs.
We introduce the holistic efficiency-accuracy driven model design strategy for YOLOs.
arXiv Detail & Related papers (2024-05-23T11:44:29Z) - YOLO-World: Real-Time Open-Vocabulary Object Detection [87.08732047660058]
We introduce YOLO-World, an innovative approach that enhances YOLO with open-vocabulary detection capabilities.
Our method excels in detecting a wide range of objects in a zero-shot manner with high efficiency.
YOLO-World achieves 35.4 AP with 52.0 FPS on V100, which outperforms many state-of-the-art methods in terms of both accuracy and speed.
arXiv Detail & Related papers (2024-01-30T18:59:38Z) - Hy-Tracker: A Novel Framework for Enhancing Efficiency and Accuracy of
Object Tracking in Hyperspectral Videos [19.733925664613093]
We propose a novel framework called Hy-Tracker to bridge the gap between hyperspectral data and state-of-the-art object detection methods.
The framework incorporates a refined tracking module on top of YOLOv7.
The experimental results on hyperspectral benchmark datasets demonstrate the effectiveness of Hy-Tracker.
arXiv Detail & Related papers (2023-11-30T02:38:45Z) - YOLO-MS: Rethinking Multi-Scale Representation Learning for Real-time
Object Detection [80.11152626362109]
We provide an efficient and performant object detector, termed YOLO-MS.
We train our YOLO-MS on the MS COCO dataset from scratch without relying on any other large-scale datasets.
Our work can also be used as a plug-and-play module for other YOLO models.
arXiv Detail & Related papers (2023-08-10T10:12:27Z) - YOLOv3 with Spatial Pyramid Pooling for Object Detection with Unmanned
Aerial Vehicles [0.0]
We aim to improve the performance of the one-stage detector YOLOv3 by adding a Spatial Pyramid Pooling layer on the end of the backbone darknet-53.
We also conducted an evaluation study on different versions of YOLOv3 methods.
arXiv Detail & Related papers (2023-05-21T04:41:52Z) - Benchmarking YOLOv5 and YOLOv7 models with DeepSORT for droplet tracking
applications [0.0]
This work is a benchmark study for the YOLOv5 and YOLOv7 networks with DeepSORT in terms of the training time and inference time for a custom dataset of microfluidic droplets.
We compare the performance of the droplet tracking applications with YOLOv5 and YOLOv7 in terms of training time and time to analyze a given video across various hardware configurations.
arXiv Detail & Related papers (2023-01-19T17:37:40Z) - A lightweight and accurate YOLO-like network for small target detection
in Aerial Imagery [94.78943497436492]
We present YOLO-S, a simple, fast and efficient network for small target detection.
YOLO-S exploits a small feature extractor based on Darknet20, as well as skip connection, via both bypass and concatenation.
YOLO-S has an 87% decrease of parameter size and almost one half FLOPs of YOLOv3, making practical the deployment for low-power industrial applications.
arXiv Detail & Related papers (2022-04-05T16:29:49Z) - Detecting Invisible People [58.49425715635312]
We re-purpose tracking benchmarks and propose new metrics for the task of detecting invisible objects.
We demonstrate that current detection and tracking systems perform dramatically worse on this task.
Second, we build dynamic models that explicitly reason in 3D, making use of observations produced by state-of-the-art monocular depth estimation networks.
arXiv Detail & Related papers (2020-12-15T16:54:45Z) - Progressive Object Transfer Detection [84.48927705173494]
We propose a novel Progressive Object Transfer Detection (POTD) framework.
First, POTD can leverage various object supervision of different domains effectively into a progressive detection procedure.
Second, POTD consists of two delicate transfer stages, i.e., Low-Shot Transfer Detection (LSTD), and Weakly-Supervised Transfer Detection (WSTD)
arXiv Detail & Related papers (2020-02-12T00:16:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.