Low-cost modular devices for on-road vehicle detection and characterisation
- URL: http://arxiv.org/abs/2405.00678v1
- Date: Fri, 26 Jan 2024 16:42:51 GMT
- Title: Low-cost modular devices for on-road vehicle detection and characterisation
- Authors: Jose-Luis Poza-Lujan, Pedro Uribe-Chavert, Juan-Luis Posadas-Yagüe,
- Abstract summary: This article introduces a system based on modular devices that is economical and has a low computational cost.
The devices use ultrasonic sensors to detect the speed and length of vehicles.
The measurement accuracy is improved through the collaboration of the device modules.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Detecting and characterising vehicles is one of the purposes of embedded systems used in intelligent environments. An analysis of a vehicle characteristics can reveal inappropriate or dangerous behaviour. This detection makes it possible to sanction or notify emergency services to take early and practical actions. Vehicle detection and characterisation systems employ complex sensors such as video cameras, especially in urban environments. These sensors provide high precision and performance, although the price and computational requirements are proportional to their accuracy. These sensors offer high accuracy, but the price and computational requirements are directly proportional to their performance. This article introduces a system based on modular devices that is economical and has a low computational cost. These devices use ultrasonic sensors to detect the speed and length of vehicles. The measurement accuracy is improved through the collaboration of the device modules. The experiments were performed using multiple modules oriented to different angles. This module is coupled with another specifically designed to detect distance using previous modules speed and length data. The collaboration between different modules reduces the speed relative error ranges from 1 to 5, depending on the angle configuration used in the modules.
Related papers
- Learning 3D Perception from Others' Predictions [64.09115694891679]
We investigate a new scenario to construct 3D object detectors: learning from the predictions of a nearby unit that is equipped with an accurate detector.
For example, when a self-driving car enters a new area, it may learn from other traffic participants whose detectors have been optimized for that area.
arXiv Detail & Related papers (2024-10-03T16:31:28Z) - Automated Automotive Radar Calibration With Intelligent Vehicles [73.15674960230625]
We present an approach for automated and geo-referenced calibration of automotive radar sensors.
Our method does not require external modifications of a vehicle and instead uses the location data obtained from automated vehicles.
Our evaluation on data from a real testing site shows that our method can correctly calibrate infrastructure sensors in an automated manner.
arXiv Detail & Related papers (2023-06-23T07:01:10Z) - Energy-Based Models for Cross-Modal Localization using Convolutional
Transformers [52.27061799824835]
We present a novel framework for localizing a ground vehicle mounted with a range sensor against satellite imagery in the absence of GPS.
We propose a method using convolutional transformers that performs accurate metric-level localization in a cross-modal manner.
We train our model end-to-end and demonstrate our approach achieving higher accuracy than the state-of-the-art on KITTI, Pandaset, and a custom dataset.
arXiv Detail & Related papers (2023-06-06T21:27:08Z) - Deep Learning for Inertial Sensor Alignment [1.9773109138840514]
We propose a data-driven approach to learn the yaw mounting angle of a smartphone equipped with an inertial measurement unit (IMU) and strapped to a car.
The proposed model uses only the accelerometer and gyroscope readings from an IMU as input.
The trained model is deployed on an Android device and evaluated in real-time to test the accuracy of the estimated yaw mounting angle.
arXiv Detail & Related papers (2022-12-10T07:50:29Z) - Extrinsic Camera Calibration with Semantic Segmentation [60.330549990863624]
We present an extrinsic camera calibration approach that automatizes the parameter estimation by utilizing semantic segmentation information.
Our approach relies on a coarse initial measurement of the camera pose and builds on lidar sensors mounted on a vehicle.
We evaluate our method on simulated and real-world data to demonstrate low error measurements in the calibration results.
arXiv Detail & Related papers (2022-08-08T07:25:03Z) - Anomaly Detection and Inter-Sensor Transfer Learning on Smart
Manufacturing Datasets [6.114996271792091]
In many cases, the goal of the smart manufacturing system is to rapidly detect (or anticipate) failures to reduce operational cost and eliminate downtime.
This often boils down to detecting anomalies within the sensor date acquired from the system.
The smart manufacturing application domain poses certain salient technical challenges.
We show that predictive failure classification can be achieved, thus paving the way for predictive maintenance.
arXiv Detail & Related papers (2022-06-13T17:51:24Z) - SenseMag: Enabling Low-Cost Traffic Monitoring using Non-invasive
Magnetic Sensing [40.55828486246774]
This paper introduces a low-cost method, namely SenseMag, to recognize the vehicular type using a pair of non-invasive magnetic sensors deployed on the straight road section.
SenseMag filters out noises and segments received magnetic signals by the exact time points that the vehicle arrives or departs from every sensor node.
Our field experiment results validate that SenseMag is with at least $90%$ vehicle type classification accuracy and less than 5% vehicle length classification error.
arXiv Detail & Related papers (2021-10-24T07:47:43Z) - On the Role of Sensor Fusion for Object Detection in Future Vehicular
Networks [25.838878314196375]
We evaluate how using a combination of different sensors affects the detection of the environment in which the vehicles move and operate.
The final objective is to identify the optimal setup that would minimize the amount of data to be distributed over the channel.
arXiv Detail & Related papers (2021-04-23T18:58:37Z) - Automatic Extrinsic Calibration Method for LiDAR and Camera Sensor
Setups [68.8204255655161]
We present a method to calibrate the parameters of any pair of sensors involving LiDARs, monocular or stereo cameras.
The proposed approach can handle devices with very different resolutions and poses, as usually found in vehicle setups.
arXiv Detail & Related papers (2021-01-12T12:02:26Z) - Investigating the Effect of Sensor Modalities in Multi-Sensor
Detection-Prediction Models [8.354898936252516]
We focus on the contribution of sensor modalities towards the model performance.
In addition, we investigate the use of sensor dropout to mitigate the above-mentioned issues.
arXiv Detail & Related papers (2021-01-09T03:21:36Z) - Deep Soft Procrustes for Markerless Volumetric Sensor Alignment [81.13055566952221]
In this work, we improve markerless data-driven correspondence estimation to achieve more robust multi-sensor spatial alignment.
We incorporate geometric constraints in an end-to-end manner into a typical segmentation based model and bridge the intermediate dense classification task with the targeted pose estimation one.
Our model is experimentally shown to achieve similar results with marker-based methods and outperform the markerless ones, while also being robust to the pose variations of the calibration structure.
arXiv Detail & Related papers (2020-03-23T10:51:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.