MonoPIC -- A Monocular Low-Latency Pedestrian Intention Classification
Framework for IoT Edges Using ID3 Modelled Decision Trees
- URL: http://arxiv.org/abs/2304.00206v3
- Date: Sun, 4 Feb 2024 03:53:34 GMT
- Title: MonoPIC -- A Monocular Low-Latency Pedestrian Intention Classification
Framework for IoT Edges Using ID3 Modelled Decision Trees
- Authors: Sriram Radhakrishna, Adithya Balasubramanyam
- Abstract summary: We propose an algorithm that classifies the intent of a single arbitrarily chosen pedestrian in a two dimensional frame into logic states.
This bypasses the need to employ any relatively high latency deep-learning algorithms.
The model was able to achieve an average testing accuracy of 83.56% with a reliable variance of 0.0042 while operating with an average latency of 48 milliseconds.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Road accidents involving autonomous vehicles commonly occur in situations
where a (pedestrian) obstacle presents itself in the path of the moving vehicle
at very sudden time intervals, leaving the robot even lesser time to react to
the change in scene. In order to tackle this issue, we propose a novel
algorithmic implementation that classifies the intent of a single arbitrarily
chosen pedestrian in a two dimensional frame into logic states in a procedural
manner using quaternions generated from a MediaPipe pose estimation model. This
bypasses the need to employ any relatively high latency deep-learning
algorithms primarily due to the lack of necessity for depth perception as well
as an implicit cap on the computational resources that most IoT edge devices
present. The model was able to achieve an average testing accuracy of 83.56%
with a reliable variance of 0.0042 while operating with an average latency of
48 milliseconds, demonstrating multiple notable advantages over the current
standard of using spatio-temporal convolutional networks for these perceptive
tasks.
Related papers
- VARADE: a Variational-based AutoRegressive model for Anomaly Detection on the Edge [7.4646496981460855]
This work presents a novel solution implementing a light autoregressive framework based on variational inference, which is best suited for real-time execution on the edge.
The proposed approach was validated on a robotic arm, part of a pilot production line, and compared with several state-of-the-art algorithms.
arXiv Detail & Related papers (2024-09-23T08:46:15Z) - OPUS: Occupancy Prediction Using a Sparse Set [64.60854562502523]
We present a framework to simultaneously predict occupied locations and classes using a set of learnable queries.
OPUS incorporates a suite of non-trivial strategies to enhance model performance.
Our lightest model achieves superior RayIoU on the Occ3D-nuScenes dataset at near 2x FPS, while our heaviest model surpasses previous best results by 6.1 RayIoU.
arXiv Detail & Related papers (2024-09-14T07:44:22Z) - Event-Aided Time-to-Collision Estimation for Autonomous Driving [28.13397992839372]
We present a novel method that estimates the time to collision using a neuromorphic event-based camera.
The proposed algorithm consists of a two-step approach for efficient and accurate geometric model fitting on event data.
Experiments on both synthetic and real data demonstrate the effectiveness of the proposed method.
arXiv Detail & Related papers (2024-07-10T02:37:36Z) - oTTC: Object Time-to-Contact for Motion Estimation in Autonomous Driving [4.707950656037167]
autonomous driving systems rely heavily on object detection to avoid collisions and drive safely.
Monocular 3D object detectors try to solve this problem by directly predicting 3D bounding boxes and object velocities given a camera image.
Recent research estimates time-to-contact in a per-pixel manner and suggests that it is more effective measure than velocity and depth combined.
We propose per-object time-to-contact estimation by extending object detection models to additionally predict the time-to-contact attribute for each object.
arXiv Detail & Related papers (2024-05-13T12:34:18Z) - Rethinking Voxelization and Classification for 3D Object Detection [68.8204255655161]
The main challenge in 3D object detection from LiDAR point clouds is achieving real-time performance without affecting the reliability of the network.
We present a solution to improve network inference speed and precision at the same time by implementing a fast dynamic voxelizer.
In addition, we propose a lightweight detection sub-head model for classifying predicted objects and filter out false detected objects.
arXiv Detail & Related papers (2023-01-10T16:22:04Z) - Correlating sparse sensing for large-scale traffic speed estimation: A
Laplacian-enhanced low-rank tensor kriging approach [76.45949280328838]
We propose a Laplacian enhanced low-rank tensor (LETC) framework featuring both lowrankness and multi-temporal correlations for large-scale traffic speed kriging.
We then design an efficient solution algorithm via several effective numeric techniques to scale up the proposed model to network-wide kriging.
arXiv Detail & Related papers (2022-10-21T07:25:57Z) - Achieving Real-Time Object Detection on MobileDevices with Neural
Pruning Search [45.20331644857981]
We propose a compiler-aware neural pruning search framework to achieve high-speed inference on autonomous vehicles for 2D and 3D object detection.
For the first time, the proposed method achieves computation (close-to) real-time, 55ms and 99ms inference times for YOLOv4 based 2D object detection and PointPillars based 3D detection.
arXiv Detail & Related papers (2021-06-28T18:59:20Z) - ANNETTE: Accurate Neural Network Execution Time Estimation with Stacked
Models [56.21470608621633]
We propose a time estimation framework to decouple the architectural search from the target hardware.
The proposed methodology extracts a set of models from micro- kernel and multi-layer benchmarks and generates a stacked model for mapping and network execution time estimation.
We compare estimation accuracy and fidelity of the generated mixed models, statistical models with the roofline model, and a refined roofline model for evaluation.
arXiv Detail & Related papers (2021-05-07T11:39:05Z) - BEVDetNet: Bird's Eye View LiDAR Point Cloud based Real-time 3D Object
Detection for Autonomous Driving [6.389322215324224]
We propose a novel semantic segmentation architecture as a single unified model for object center detection using key points, box predictions and orientation prediction.
The proposed architecture can be trivially extended to include semantic segmentation classes like road without any additional computation.
The model is 5X faster than other top accuracy models with a minimal accuracy degradation of 2% in Average Precision at IoU=0.5 on KITTI dataset.
arXiv Detail & Related papers (2021-04-21T22:06:39Z) - Multi-scale Interaction for Real-time LiDAR Data Segmentation on an
Embedded Platform [62.91011959772665]
Real-time semantic segmentation of LiDAR data is crucial for autonomously driving vehicles.
Current approaches that operate directly on the point cloud use complex spatial aggregation operations.
We propose a projection-based method, called Multi-scale Interaction Network (MINet), which is very efficient and accurate.
arXiv Detail & Related papers (2020-08-20T19:06:11Z) - Risk-Averse MPC via Visual-Inertial Input and Recurrent Networks for
Online Collision Avoidance [95.86944752753564]
We propose an online path planning architecture that extends the model predictive control (MPC) formulation to consider future location uncertainties.
Our algorithm combines an object detection pipeline with a recurrent neural network (RNN) which infers the covariance of state estimates.
The robustness of our methods is validated on complex quadruped robot dynamics and can be generally applied to most robotic platforms.
arXiv Detail & Related papers (2020-07-28T07:34:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.