Real-time HOG+SVM based object detection using SoC FPGA for a UHD video
stream
- URL: http://arxiv.org/abs/2204.10619v1
- Date: Fri, 22 Apr 2022 10:29:21 GMT
- Title: Real-time HOG+SVM based object detection using SoC FPGA for a UHD video
stream
- Authors: Mateusz Wasala and Tomasz Kryjak
- Abstract summary: We present a real-time implementation of the well-known pedestrian detector with HOG (Histogram of Oriented Gradients) feature extraction and SVM (Support Vector Machine) classification.
The system is capable of detecting a pedestrian in a single scale.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Object detection is an essential component of many vision systems. For
example, pedestrian detection is used in advanced driver assistance systems
(ADAS) and advanced video surveillance systems (AVSS). Currently, most
detectors use deep convolutional neural networks (e.g., the YOLO -- You Only
Look Once -- family), which, however, due to their high computational
complexity, are not able to process a very high-resolution video stream in
real-time, especially within a limited energy budget. In this paper we present
a hardware implementation of the well-known pedestrian detector with HOG
(Histogram of Oriented Gradients) feature extraction and SVM (Support Vector
Machine) classification. Our system running on AMD Xilinx Zynq UltraScale+
MPSoC (Multiprocessor System on Chip) device allows real-time processing of 4K
resolution (UHD -- Ultra High Definition, 3840 x 2160 pixels) video for 60
frames per second. The system is capable of detecting a pedestrian in a single
scale. The results obtained confirm the high suitability of reprogrammable
devices in the real-time implementation of embedded vision systems.
Related papers
- Benchmarking Jetson Edge Devices with an End-to-end Video-based Anomaly
Detection System [0.0]
We implement an end-to-end video-based crime-scene anomaly detection system inputting from surveillance videos.
The system is deployed and operates on multiple Jetson edge devices (Nano, AGX Xavier, Orin Nano)
We provide the experience of an AI-based system deployment on various Jetson Edge devices with Docker technology.
arXiv Detail & Related papers (2023-07-28T17:16:57Z) - Agile gesture recognition for capacitive sensing devices: adapting
on-the-job [55.40855017016652]
We demonstrate a hand gesture recognition system that uses signals from capacitive sensors embedded into the etee hand controller.
The controller generates real-time signals from each of the wearer five fingers.
We use a machine learning technique to analyse the time series signals and identify three features that can represent 5 fingers within 500 ms.
arXiv Detail & Related papers (2023-05-12T17:24:02Z) - Deep Learning Computer Vision Algorithms for Real-time UAVs On-board
Camera Image Processing [77.34726150561087]
This paper describes how advanced deep learning based computer vision algorithms are applied to enable real-time on-board sensor processing for small UAVs.
All algorithms have been developed using state-of-the-art image processing methods based on deep neural networks.
arXiv Detail & Related papers (2022-11-02T11:10:42Z) - ETAD: A Unified Framework for Efficient Temporal Action Detection [70.21104995731085]
Untrimmed video understanding such as temporal action detection (TAD) often suffers from the pain of huge demand for computing resources.
We build a unified framework for efficient end-to-end temporal action detection (ETAD)
ETAD achieves state-of-the-art performance on both THUMOS-14 and ActivityNet-1.3.
arXiv Detail & Related papers (2022-05-14T21:16:21Z) - E^2TAD: An Energy-Efficient Tracking-based Action Detector [78.90585878925545]
This paper presents a tracking-based solution to accurately and efficiently localize predefined key actions.
It won first place in the UAV-Video Track of 2021 Low-Power Computer Vision Challenge (LPCVC)
arXiv Detail & Related papers (2022-04-09T07:52:11Z) - Argus++: Robust Real-time Activity Detection for Unconstrained Video
Streams with Overlapping Cube Proposals [85.76513755331318]
Argus++ is a robust real-time activity detection system for analyzing unconstrained video streams.
The overall system is optimized for real-time processing on standalone consumer-level hardware.
arXiv Detail & Related papers (2022-01-14T03:35:22Z) - Parallel Detection for Efficient Video Analytics at the Edge [5.547133811014004]
Deep Neural Network (DNN) trained object detectors are widely deployed in mission-critical systems for real time video analytics at the edge.
A common performance requirement in mission-critical edge services is the near real-time latency of online object detection on edge devices.
This paper addresses these problems by exploiting multi-model multi-device detection parallelism for fast object detection in edge systems.
arXiv Detail & Related papers (2021-07-27T02:50:46Z) - High Performance Hyperspectral Image Classification using Graphics
Processing Units [0.0]
Real-time remote sensing applications require onboard real time processing capabilities.
Lightweight, small size and low power consumption hardware is essential for onboard real time processing systems.
arXiv Detail & Related papers (2021-05-30T09:26:03Z) - Motion Vector Extrapolation for Video Object Detection [0.0]
MOVEX enables low latency video object detection on common CPU based systems.
We show that our approach significantly reduces the baseline latency of any given object detector.
Further latency reduction, up to 25x lower than the original latency, can be achieved with minimal accuracy loss.
arXiv Detail & Related papers (2021-04-18T17:26:37Z) - ACDnet: An action detection network for real-time edge computing based
on flow-guided feature approximation and memory aggregation [8.013823319651395]
ACDnet is a compact action detection network targeting real-time edge computing.
It exploits the temporal coherence between successive video frames to approximate CNN features rather than naively extracting them.
It can robustly achieve detection well above real-time (75 FPS)
arXiv Detail & Related papers (2021-02-26T14:06:31Z) - Single Shot Video Object Detector [215.06904478667337]
Single Shot Video Object Detector (SSVD) is a new architecture that novelly integrates feature aggregation into a one-stage detector for object detection in videos.
For $448 times 448$ input, SSVD achieves 79.2% mAP on ImageNet VID dataset.
arXiv Detail & Related papers (2020-07-07T15:36:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.