Dense Residual Networks for Gaze Mapping on Indian Roads
- URL: http://arxiv.org/abs/2203.11611v1
- Date: Tue, 22 Mar 2022 10:58:02 GMT
- Title: Dense Residual Networks for Gaze Mapping on Indian Roads
- Authors: Chaitanya Kapoor, Kshitij Kumar, Soumya Vishnoi and Sriram Ramanathan
- Abstract summary: We propose a novel architecture, DR-Gaze, which is used to map the driver's gaze onto the road.
We compare our results with previous works and state-of-the-art results on the DGAZE dataset.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In the recent past, greater accessibility to powerful computational resources
has enabled progress in the field of Deep Learning and Computer Vision to grow
by leaps and bounds. This in consequence has lent progress to the domain of
Autonomous Driving and Navigation Systems. Most of the present research work
has been focused on driving scenarios in the European or American roads. Our
paper draws special attention to the Indian driving context. To this effect, we
propose a novel architecture, DR-Gaze, which is used to map the driver's gaze
onto the road. We compare our results with previous works and state-of-the-art
results on the DGAZE dataset. Our code will be made publicly available upon
acceptance of our paper.
Related papers
- Spatial Retrieval Augmented Autonomous Driving [81.39665750557526]
Existing autonomous driving systems rely on onboard sensors for environmental perception.<n>We propose the spatial retrieval paradigm, introducing offline retrieved geographic images as an additional input.<n>We will open-source dataset curation code, data, and benchmarks for further study of this new autonomous driving paradigm.
arXiv Detail & Related papers (2025-12-07T14:40:49Z) - nuScenes Revisited: Progress and Challenges in Autonomous Driving [12.38901090939622]
We revisit one of the most widely used autonomous driving datasets: the nuScenes dataset.<n> nuScenes exemplifies key trends in AV development, being the first dataset to include radar data.<n>We provide an unprecedented look into the creation of nuScenes, as well as its extensions nuImages and Panoptic nuScenes.
arXiv Detail & Related papers (2025-12-02T06:14:28Z) - Solving Scene Understanding for Autonomous Navigation in Unstructured Environments [0.0]
This paper performs semantic segmentation on the Indian Driving dataset.<n>The dataset is more challenging compared to other datasets like Cityscapes.<n>Five different models have been trained and their performance has been compared using the Mean Intersection over Union.
arXiv Detail & Related papers (2025-07-27T19:11:21Z) - The Role of World Models in Shaping Autonomous Driving: A Comprehensive Survey [50.62538723793247]
Driving World Model (DWM) focuses on predicting scene evolution during the driving process.
DWM methods enable autonomous driving systems to better perceive, understand, and interact with dynamic driving environments.
arXiv Detail & Related papers (2025-02-14T18:43:15Z) - Parking Space Detection in the City of Granada [0.0]
This paper addresses the challenge of parking space detection in urban areas, focusing on the city of Granada.
We develop and apply semantic segmentation techniques to accurately identify parked cars, moving cars and roads.
We employ Fully Convolutional Networks, Pyramid Networks and Dilated Convolutions, demonstrating their effectiveness in urban semantic segmentation.
arXiv Detail & Related papers (2025-01-11T22:29:12Z) - Monocular Lane Detection Based on Deep Learning: A Survey [51.19079381823076]
Lane detection plays an important role in autonomous driving perception systems.
As deep learning algorithms gain popularity, monocular lane detection methods based on deep learning have demonstrated superior performance.
This paper presents a comprehensive overview of existing methods, encompassing both the increasingly mature 2D lane detection approaches and the developing 3D lane detection works.
arXiv Detail & Related papers (2024-11-25T12:09:43Z) - Leveraging GNSS and Onboard Visual Data from Consumer Vehicles for Robust Road Network Estimation [18.236615392921273]
This paper addresses the challenge of road graph construction for autonomous vehicles.
We propose using global navigation satellite system (GNSS) traces and basic image data acquired from these standard sensors in consumer vehicles.
We exploit the spatial information in the data by framing the problem as a road centerline semantic segmentation task using a convolutional neural network.
arXiv Detail & Related papers (2024-08-03T02:57:37Z) - Enhancing End-to-End Autonomous Driving with Latent World Model [78.22157677787239]
We propose a novel self-supervised learning approach using the LAtent World model (LAW) for end-to-end driving.
LAW predicts future scene features based on current features and ego trajectories.
This self-supervised task can be seamlessly integrated into perception-free and perception-based frameworks.
arXiv Detail & Related papers (2024-06-12T17:59:21Z) - Networking Systems for Video Anomaly Detection: A Tutorial and Survey [55.28514053969056]
Video Anomaly Detection (VAD) is a fundamental research task within the Artificial Intelligence (AI) community.
In this article, we delineate the foundational assumptions, learning frameworks, and applicable scenarios of various deep learning-driven VAD routes.
We showcase our latest NSVAD research in industrial IoT and smart cities, along with an end-cloud collaborative architecture for deployable NSVAD.
arXiv Detail & Related papers (2024-05-16T02:00:44Z) - RAG-based Explainable Prediction of Road Users Behaviors for Automated Driving using Knowledge Graphs and Large Language Models [8.253092044813595]
We propose an explainable road users' behavior prediction system that integrates the reasoning abilities of Knowledge Graphs and Large Language Models.
Two use cases have been implemented following the proposed approach: 1) Prediction of pedestrians' crossing actions; 2) Prediction of lane change maneuvers.
arXiv Detail & Related papers (2024-05-01T11:06:31Z) - G-MEMP: Gaze-Enhanced Multimodal Ego-Motion Prediction in Driving [71.9040410238973]
We focus on inferring the ego trajectory of a driver's vehicle using their gaze data.
Next, we develop G-MEMP, a novel multimodal ego-trajectory prediction network that combines GPS and video input with gaze data.
The results show that G-MEMP significantly outperforms state-of-the-art methods in both benchmarks.
arXiv Detail & Related papers (2023-12-13T23:06:30Z) - A Big-Data Driven Framework to Estimating Vehicle Volume based on Mobile
Device Location Data [0.40631409309544836]
Vehicle volume serves as a critical metric and the fundamental basis for traffic signal control, transportation project prioritization, road maintenance and more.
Traditional methods of quantifying vehicle volume rely on manual counting, video cameras, and loop detectors at a limited number of locations.
This paper presents a big-data driven framework that can ingest terabytes of Device Location Data and estimate vehicle volume at a larger geographical area with a larger sample size.
arXiv Detail & Related papers (2023-01-20T16:23:17Z) - Sensing and Mapping for Better Roads: Initial Plan for Using Federated
Learning and Implementing a Digital Twin to Identify the Road Conditions in a
Developing Country -- Sri Lanka [0.0]
We propose how a developing country like Sri Lanka can benefit from privacy-enabled machine learning techniques.
We propose the idea of implementing a Digital Twin for the national road system in Sri Lanka.
arXiv Detail & Related papers (2021-07-30T11:06:32Z) - Embedded Vision for Self-Driving on Forest Roads [0.0]
AMTU is a robotic system designed to autonomously navigate off-road terrain and inspect if any deforestation or damage occurred along tracked route.
AMTU's core component is its embedded vision module, optimized for real-time environment perception.
We show experimental results on the test track of our research facility.
arXiv Detail & Related papers (2021-05-27T09:05:08Z) - Injecting Knowledge in Data-driven Vehicle Trajectory Predictors [82.91398970736391]
Vehicle trajectory prediction tasks have been commonly tackled from two perspectives: knowledge-driven or data-driven.
In this paper, we propose to learn a "Realistic Residual Block" (RRB) which effectively connects these two perspectives.
Our proposed method outputs realistic predictions by confining the residual range and taking into account its uncertainty.
arXiv Detail & Related papers (2021-03-08T16:03:09Z) - Occupancy Anticipation for Efficient Exploration and Navigation [97.17517060585875]
We propose occupancy anticipation, where the agent uses its egocentric RGB-D observations to infer the occupancy state beyond the visible regions.
By exploiting context in both the egocentric views and top-down maps our model successfully anticipates a broader map of the environment.
Our approach is the winning entry in the 2020 Habitat PointNav Challenge.
arXiv Detail & Related papers (2020-08-21T03:16:51Z) - Learning Accurate and Human-Like Driving using Semantic Maps and
Attention [152.48143666881418]
This paper investigates how end-to-end driving models can be improved to drive more accurately and human-like.
We exploit semantic and visual maps from HERE Technologies and augment the existing Drive360 dataset with such.
Our models are trained and evaluated on the Drive360 + HERE dataset, which features 60 hours and 3000 km of real-world driving data.
arXiv Detail & Related papers (2020-07-10T22:25:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.