Deep Learning Enhanced Road Traffic Analysis: Scalable Vehicle Detection and Velocity Estimation Using PlanetScope Imagery
- URL: http://arxiv.org/abs/2410.14698v1
- Date: Fri, 04 Oct 2024 18:14:07 GMT
- Title: Deep Learning Enhanced Road Traffic Analysis: Scalable Vehicle Detection and Velocity Estimation Using PlanetScope Imagery
- Authors: Maciej Adamiak, Yulia Grinblat, Julian Psotta, Nir Fulman, Himshikhar Mazumdar, Shiyu Tang, Alexander Zipf,
- Abstract summary: This paper presents a method for detecting and estimating vehicle speeds using PlanetScope SuperDove satellite imagery.
We propose a Keypoint R-CNN model to track vehicle trajectories across RGB bands, leveraging band timing differences to estimate speed.
Results from drone comparison reveal underestimations, with average speeds of 112.85 km/h for satellite data versus 131.83 km/h from drone footage.
- Score: 38.22365259129059
- License:
- Abstract: This paper presents a method for detecting and estimating vehicle speeds using PlanetScope SuperDove satellite imagery, offering a scalable solution for global vehicle traffic monitoring. Conventional methods such as stationary sensors and mobile systems like UAVs are limited in coverage and constrained by high costs and legal restrictions. Satellite-based approaches provide broad spatial coverage but face challenges, including high costs, low frame rates, and difficulty detecting small vehicles in high-resolution imagery. We propose a Keypoint R-CNN model to track vehicle trajectories across RGB bands, leveraging band timing differences to estimate speed. Validation is performed using drone footage and GPS data covering highways in Germany and Poland. Our model achieved a Mean Average Precision of 0.53 and velocity estimation errors of approximately 3.4 m/s compared to GPS data. Results from drone comparison reveal underestimations, with average speeds of 112.85 km/h for satellite data versus 131.83 km/h from drone footage. While challenges remain with high-speed accuracy, this approach demonstrates the potential for scalable, daily traffic monitoring across vast areas, providing valuable insights into global traffic dynamics.
Related papers
- Advanced computer vision for extracting georeferenced vehicle trajectories from drone imagery [4.387337528923525]
This paper presents a framework for extracting georeferenced vehicle trajectories from high-altitude drone footage.
We employ state-of-the-art computer vision and deep learning to create an end-to-end pipeline.
Results demonstrate the potential of integrating drone technology with advanced computer vision for precise, cost-effective urban traffic monitoring.
arXiv Detail & Related papers (2024-11-04T14:49:01Z) - RoadRunner -- Learning Traversability Estimation for Autonomous Off-road Driving [13.101416329887755]
We present RoadRunner, a framework capable of predicting terrain traversability and an elevation map directly from camera and LiDAR sensor inputs.
RoadRunner enables reliable autonomous navigation, by fusing sensory information, handling of uncertainty, and generation of contextually informed predictions.
We demonstrate the effectiveness of RoadRunner in enabling safe and reliable off-road navigation at high speeds in multiple real-world driving scenarios through unstructured desert environments.
arXiv Detail & Related papers (2024-02-29T16:47:54Z) - Vehicle Perception from Satellite [54.07157185000604]
The dataset is constructed based on 12 satellite videos and 14 synthetic videos recorded from GTA-V.
It supports several tasks, including tiny object detection, counting and density estimation.
128,801 vehicles are annotated totally, and the number of vehicles in each image varies from 0 to 101.
arXiv Detail & Related papers (2024-02-01T15:59:16Z) - FARSEC: A Reproducible Framework for Automatic Real-Time Vehicle Speed
Estimation Using Traffic Cameras [14.339217121537537]
Transportation-dependent systems, such as for navigation and logistics, have great potential to benefit from reliable speed estimation.
We provide a novel framework for automatic real-time vehicle speed calculation, which copes with more diverse data from publicly available traffic cameras.
Our framework is capable of handling realistic conditions such as camera movements and different video stream inputs automatically.
arXiv Detail & Related papers (2023-09-25T19:02:40Z) - Turning Traffic Monitoring Cameras into Intelligent Sensors for Traffic
Density Estimation [9.096163152559054]
This paper proposes a framework for estimating traffic density using uncalibrated traffic monitoring cameras with 4L characteristics.
The proposed framework consists of two major components: camera calibration and vehicle detection.
The results show that the Mean Absolute Error (MAE) in camera calibration is less than 0.2 meters out of 6 meters, and the accuracy of vehicle detection under various conditions is approximately 90%.
arXiv Detail & Related papers (2021-10-29T15:39:06Z) - Real Time Monocular Vehicle Velocity Estimation using Synthetic Data [78.85123603488664]
We look at the problem of estimating the velocity of road vehicles from a camera mounted on a moving car.
We propose a two-step approach where first an off-the-shelf tracker is used to extract vehicle bounding boxes and then a small neural network is used to regress the vehicle velocity.
arXiv Detail & Related papers (2021-09-16T13:10:27Z) - R4Dyn: Exploring Radar for Self-Supervised Monocular Depth Estimation of
Dynamic Scenes [69.6715406227469]
Self-supervised monocular depth estimation in driving scenarios has achieved comparable performance to supervised approaches.
We present R4Dyn, a novel set of techniques to use cost-efficient radar data on top of a self-supervised depth estimation framework.
arXiv Detail & Related papers (2021-08-10T17:57:03Z) - Vision-based Vehicle Speed Estimation for ITS: A Survey [0.47248250311484113]
The number of speed cameras installed worldwide has been growing in recent years.
Traffic monitoring and forecasting in road networks plays a fundamental role to enhance traffic, emissions and energy consumption in smart cities.
The use of vision-based systems brings great challenges to be solved, but also great potential advantages.
arXiv Detail & Related papers (2021-01-15T15:07:54Z) - LiRaNet: End-to-End Trajectory Prediction using Spatio-Temporal Radar
Fusion [52.59664614744447]
We present LiRaNet, a novel end-to-end trajectory prediction method which utilizes radar sensor information along with widely used lidar and high definition (HD) maps.
automotive radar provides rich, complementary information, allowing for longer range vehicle detection as well as instantaneous velocity measurements.
arXiv Detail & Related papers (2020-10-02T00:13:00Z) - Drone-based RGB-Infrared Cross-Modality Vehicle Detection via
Uncertainty-Aware Learning [59.19469551774703]
Drone-based vehicle detection aims at finding the vehicle locations and categories in an aerial image.
We construct a large-scale drone-based RGB-Infrared vehicle detection dataset, termed DroneVehicle.
Our DroneVehicle collects 28, 439 RGB-Infrared image pairs, covering urban roads, residential areas, parking lots, and other scenarios from day to night.
arXiv Detail & Related papers (2020-03-05T05:29:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.