EgoSpeed-Net: Forecasting Speed-Control in Driver Behavior from
Egocentric Video Data
- URL: http://arxiv.org/abs/2209.13459v1
- Date: Tue, 27 Sep 2022 15:25:57 GMT
- Title: EgoSpeed-Net: Forecasting Speed-Control in Driver Behavior from
Egocentric Video Data
- Authors: Yichen Ding, Ziming Zhang, Yanhua Li, Xun Zhou
- Abstract summary: We propose a novel graph convolutional network (GCN) based network, namely, EgoSpeed-Net.
We are motivated by the fact that the position changes of objects over time can provide us very useful clues for forecasting the speed change in future.
We conduct extensive experiments on the Honda Research Institute Driving dataset and demonstrate the superior performance of EgoSpeed-Net.
- Score: 24.32406053197066
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Speed-control forecasting, a challenging problem in driver behavior analysis,
aims to predict the future actions of a driver in controlling vehicle speed
such as braking or acceleration. In this paper, we try to address this
challenge solely using egocentric video data, in contrast to the majority of
works in the literature using either third-person view data or extra vehicle
sensor data such as GPS, or both. To this end, we propose a novel graph
convolutional network (GCN) based network, namely, EgoSpeed-Net. We are
motivated by the fact that the position changes of objects over time can
provide us very useful clues for forecasting the speed change in future. We
first model the spatial relations among the objects from each class, frame by
frame, using fully-connected graphs, on top of which GCNs are applied for
feature extraction. Then we utilize a long short-term memory network to fuse
such features per class over time into a vector, concatenate such vectors and
forecast a speed-control action using a multilayer perceptron classifier. We
conduct extensive experiments on the Honda Research Institute Driving Dataset
and demonstrate the superior performance of EgoSpeed-Net.
Related papers
- Object Detection in Thermal Images Using Deep Learning for Unmanned
Aerial Vehicles [0.9208007322096533]
This work presents a neural network model capable of recognizing small and tiny objects in thermal images collected by unmanned aerial vehicles.
The backbone is developed based on the structure of YOLOv5 combined with the use of a transformer encoder at the end.
The neck includes a BI-FPN block combined with the use of a sliding window and a transformer to increase the information fed into the prediction head.
arXiv Detail & Related papers (2024-02-13T06:40:55Z) - G-MEMP: Gaze-Enhanced Multimodal Ego-Motion Prediction in Driving [71.9040410238973]
We focus on inferring the ego trajectory of a driver's vehicle using their gaze data.
Next, we develop G-MEMP, a novel multimodal ego-trajectory prediction network that combines GPS and video input with gaze data.
The results show that G-MEMP significantly outperforms state-of-the-art methods in both benchmarks.
arXiv Detail & Related papers (2023-12-13T23:06:30Z) - Robust Autonomous Vehicle Pursuit without Expert Steering Labels [41.168074206046164]
We present a learning method for lateral and longitudinal motion control of an ego-vehicle for vehicle pursuit.
The car being controlled does not have a pre-defined route, rather it reactively adapts to follow a target vehicle while maintaining a safety distance.
We extensively validate our approach using the CARLA simulator on a wide range of terrains.
arXiv Detail & Related papers (2023-08-16T14:09:39Z) - Interaction-Aware Personalized Vehicle Trajectory Prediction Using
Temporal Graph Neural Networks [8.209194305630229]
Existing methods mainly rely on generic trajectory predictions from large datasets.
We propose an approach for interaction-aware personalized vehicle trajectory prediction that incorporates temporal graph neural networks.
arXiv Detail & Related papers (2023-08-14T20:20:26Z) - Motion Planning and Control for Multi Vehicle Autonomous Racing at High
Speeds [100.61456258283245]
This paper presents a multi-layer motion planning and control architecture for autonomous racing.
The proposed solution has been applied on a Dallara AV-21 racecar and tested at oval race tracks achieving lateral accelerations up to 25 $m/s2$.
arXiv Detail & Related papers (2022-07-22T15:16:54Z) - TransFollower: Long-Sequence Car-Following Trajectory Prediction through
Transformer [44.93030718234555]
We develop a long-sequence car-following trajectory prediction model based on the attention-based Transformer model.
We train and test our model with 112,597 real-world car-following events extracted from the Shanghai Naturalistic Driving Study (SH-NDS)
arXiv Detail & Related papers (2022-02-04T07:59:22Z) - Real Time Monocular Vehicle Velocity Estimation using Synthetic Data [78.85123603488664]
We look at the problem of estimating the velocity of road vehicles from a camera mounted on a moving car.
We propose a two-step approach where first an off-the-shelf tracker is used to extract vehicle bounding boxes and then a small neural network is used to regress the vehicle velocity.
arXiv Detail & Related papers (2021-09-16T13:10:27Z) - IntentNet: Learning to Predict Intention from Raw Sensor Data [86.74403297781039]
In this paper, we develop a one-stage detector and forecaster that exploits both 3D point clouds produced by a LiDAR sensor as well as dynamic maps of the environment.
Our multi-task model achieves better accuracy than the respective separate modules while saving computation, which is critical to reducing reaction time in self-driving applications.
arXiv Detail & Related papers (2021-01-20T00:31:52Z) - PnPNet: End-to-End Perception and Prediction with Tracking in the Loop [82.97006521937101]
We tackle the problem of joint perception and motion forecasting in the context of self-driving vehicles.
We propose Net, an end-to-end model that takes as input sensor data, and outputs at each time step object tracks and their future level.
arXiv Detail & Related papers (2020-05-29T17:57:25Z) - VectorNet: Encoding HD Maps and Agent Dynamics from Vectorized
Representation [74.56282712099274]
This paper introduces VectorNet, a hierarchical graph neural network that exploits the spatial locality of individual road components represented by vectors.
By operating on the vectorized high definition (HD) maps and agent trajectories, we avoid lossy rendering and computationally intensive ConvNet encoding steps.
We evaluate VectorNet on our in-house behavior prediction benchmark and the recently released Argoverse forecasting dataset.
arXiv Detail & Related papers (2020-05-08T19:07:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.