CueCAn: Cue Driven Contextual Attention For Identifying Missing Traffic
Signs on Unconstrained Roads
- URL: http://arxiv.org/abs/2303.02641v1
- Date: Sun, 5 Mar 2023 11:06:20 GMT
- Title: CueCAn: Cue Driven Contextual Attention For Identifying Missing Traffic
Signs on Unconstrained Roads
- Authors: Varun Gupta, Anbumani Subramanian, C.V. Jawahar, Rohit Saluja
- Abstract summary: Missing or non-existing object detection has been studied for locating missing curbs and estimating reasonable regions for pedestrians on road scene images.
We present the first and most challenging video dataset for missing objects, with multiple types of traffic signs for which the cues are visible without the signs in the scenes.
We train the encoder to classify the presence of traffic sign cues and then train the entire segmentation model end-to-end to localize missing traffic signs.
- Score: 26.649617412538717
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Unconstrained Asian roads often involve poor infrastructure, affecting
overall road safety. Missing traffic signs are a regular part of such roads.
Missing or non-existing object detection has been studied for locating missing
curbs and estimating reasonable regions for pedestrians on road scene images.
Such methods involve analyzing task-specific single object cues. In this paper,
we present the first and most challenging video dataset for missing objects,
with multiple types of traffic signs for which the cues are visible without the
signs in the scenes. We refer to it as the Missing Traffic Signs Video Dataset
(MTSVD). MTSVD is challenging compared to the previous works in two aspects i)
The traffic signs are generally not present in the vicinity of their cues, ii)
The traffic signs cues are diverse and unique. Also, MTSVD is the first
publicly available missing object dataset. To train the models for identifying
missing signs, we complement our dataset with 10K traffic sign tracks, with 40
percent of the traffic signs having cues visible in the scenes. For identifying
missing signs, we propose the Cue-driven Contextual Attention units (CueCAn),
which we incorporate in our model encoder. We first train the encoder to
classify the presence of traffic sign cues and then train the entire
segmentation model end-to-end to localize missing traffic signs. Quantitative
and qualitative analysis shows that CueCAn significantly improves the
performance of base models.
Related papers
- Cross-domain Few-shot In-context Learning for Enhancing Traffic Sign Recognition [49.20086587208214]
We propose a cross-domain few-shot in-context learning method based on the MLLM for enhancing traffic sign recognition.
By using description texts, our method reduces the cross-domain differences between template and real traffic signs.
Our approach requires only simple and uniform textual indications, without the need for large-scale traffic sign images and labels.
arXiv Detail & Related papers (2024-07-08T10:51:03Z) - OpenLane-V2: A Topology Reasoning Benchmark for Unified 3D HD Mapping [84.65114565766596]
We present OpenLane-V2, the first dataset on topology reasoning for traffic scene structure.
OpenLane-V2 consists of 2,000 annotated road scenes that describe traffic elements and their correlation to the lanes.
We evaluate various state-of-the-art methods, and present their quantitative and qualitative results on OpenLane-V2 to indicate future avenues for investigating topology reasoning in traffic scenes.
arXiv Detail & Related papers (2023-04-20T16:31:22Z) - Graph-based Topology Reasoning for Driving Scenes [102.35885039110057]
We present TopoNet, the first end-to-end framework capable of abstracting traffic knowledge beyond conventional perception tasks.
We evaluate TopoNet on the challenging scene understanding benchmark, OpenLane-V2.
arXiv Detail & Related papers (2023-04-11T15:23:29Z) - Traffic Scene Parsing through the TSP6K Dataset [109.69836680564616]
We introduce a specialized traffic monitoring dataset, termed TSP6K, with high-quality pixel-level and instance-level annotations.
The dataset captures more crowded traffic scenes with several times more traffic participants than the existing driving scenes.
We propose a detail refining decoder for scene parsing, which recovers the details of different semantic regions in traffic scenes.
arXiv Detail & Related papers (2023-03-06T02:05:14Z) - Salient Sign Detection In Safe Autonomous Driving: AI Which Reasons Over
Full Visual Context [2.799896314754614]
Various traffic signs in a driving scene have an unequal impact on the driver's decisions.
We construct a traffic sign detection model which emphasizes performance on salient signs.
We show that a model trained with Salience-Sensitive Focal Loss outperforms a model trained without.
arXiv Detail & Related papers (2023-01-14T01:47:09Z) - TrafficCAM: A Versatile Dataset for Traffic Flow Segmentation [9.744937939618161]
Existing traffic flow datasets have two major limitations.
They feature a limited number of classes, usually limited to one type of vehicle, and the scarcity of unlabelled data.
We introduce a new benchmark traffic flow image dataset called TrafficCAM.
arXiv Detail & Related papers (2022-11-17T16:14:38Z) - Perspective Aware Road Obstacle Detection [104.57322421897769]
We show that road obstacle detection techniques ignore the fact that, in practice, the apparent size of the obstacles decreases as their distance to the vehicle increases.
We leverage this by computing a scale map encoding the apparent size of a hypothetical object at every image location.
We then leverage this perspective map to generate training data by injecting onto the road synthetic objects whose size corresponds to the perspective foreshortening.
arXiv Detail & Related papers (2022-10-04T17:48:42Z) - METEOR: A Massive Dense & Heterogeneous Behavior Dataset for Autonomous
Driving [42.69638782267657]
We present a new and complex traffic dataset, METEOR, which captures traffic patterns in unstructured scenarios in India.
METEOR consists of more than 1000 one-minute video clips, over 2 million annotated frames with ego-vehicle trajectories, and more than 13 million bounding boxes for surrounding vehicles or traffic agents.
We use our novel dataset to evaluate the performance of object detection and behavior prediction algorithms.
arXiv Detail & Related papers (2021-09-16T01:01:55Z) - Automated Object Behavioral Feature Extraction for Potential Risk
Analysis based on Video Sensor [6.291501119156943]
Pedestrians are exposed to risk of death or serious injuries on roads, especially unsignalized crosswalks.
We propose an automated and simpler system for effectively extracting object behavioral features from video sensors deployed on the road.
This study demonstrates the potential for a network of connected video sensors to provide actionable data for smart cities.
arXiv Detail & Related papers (2021-07-08T01:11:31Z) - BoMuDANet: Unsupervised Adaptation for Visual Scene Understanding in
Unstructured Driving Environments [54.22535063244038]
We present an unsupervised adaptation approach for visual scene understanding in unstructured traffic environments.
Our method is designed for unstructured real-world scenarios with dense and heterogeneous traffic consisting of cars, trucks, two-and three-wheelers, and pedestrians.
arXiv Detail & Related papers (2020-09-22T08:25:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.