Diff-Net: Image Feature Difference based High-Definition Map Change
Detection
- URL: http://arxiv.org/abs/2107.07030v1
- Date: Wed, 14 Jul 2021 22:51:30 GMT
- Title: Diff-Net: Image Feature Difference based High-Definition Map Change
Detection
- Authors: Lei He and Shengjie Jiang and Xiaoqing Liang and Ning Wang and Shiyu
Song
- Abstract summary: Up-to-date High-Definition (HD) maps are essential for self-driving cars.
We present a deep neural network (DNN), Diff-Net, to detect changes in them.
Results demonstrate that our Diff-Net achieves better performance than the baseline methods and is ready to be integrated into a map production maintaining an up-to-date HD map.
- Score: 13.666189678747996
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Up-to-date High-Definition (HD) maps are essential for self-driving cars. To
achieve constantly updated HD maps, we present a deep neural network (DNN),
Diff-Net, to detect changes in them. Compared to traditional methods based on
object detectors, the essential design in our work is a parallel feature
difference calculation structure that infers map changes by comparing features
extracted from the camera and rasterized images. To generate these rasterized
images, we project map elements onto images in the camera view, yielding
meaningful map representations that can be consumed by a DNN accordingly. As we
formulate the change detection task as an object detection problem, we leverage
the anchor-based structure that predicts bounding boxes with different change
status categories. Furthermore, rather than relying on single frame input, we
introduce a spatio-temporal fusion module that fuses features from history
frames into the current, thus improving the overall performance. Finally, we
comprehensively validate our method's effectiveness using freshly collected
datasets. Results demonstrate that our Diff-Net achieves better performance
than the baseline methods and is ready to be integrated into a map production
pipeline maintaining an up-to-date HD map.
Related papers
- Progressive Query Refinement Framework for Bird's-Eye-View Semantic Segmentation from Surrounding Images [3.495246564946556]
We introduce the Multi-Resolution (MR) concept into Bird's-Eye-View (BEV) semantic segmentation for autonomous driving.
We propose a visual feature interaction network that promotes interactions between features across images and across feature levels.
We evaluate our model on a large-scale real-world dataset.
arXiv Detail & Related papers (2024-07-24T05:00:31Z) - Diffusion-based Data Augmentation for Object Counting Problems [62.63346162144445]
We develop a pipeline that utilizes a diffusion model to generate extensive training data.
We are the first to generate images conditioned on a location dot map with a diffusion model.
Our proposed counting loss for the diffusion model effectively minimizes the discrepancies between the location dot map and the crowd images generated.
arXiv Detail & Related papers (2024-01-25T07:28:22Z) - ScalableMap: Scalable Map Learning for Online Long-Range Vectorized HD
Map Construction [42.874195888422584]
We propose a novel end-to-end pipeline for online long-range vectorized high-definition (HD) map construction using on-board camera sensors.
We exploit the properties of map elements to improve the performance of map construction.
arXiv Detail & Related papers (2023-10-20T09:46:24Z) - InsMapper: Exploring Inner-instance Information for Vectorized HD
Mapping [41.59891369655983]
InsMapper harnesses inner-instance information for vectorized high-definition mapping through transformers.
InsMapper surpasses the previous state-of-the-art method, demonstrating its effectiveness and generality.
arXiv Detail & Related papers (2023-08-16T17:58:28Z) - DisPlacing Objects: Improving Dynamic Vehicle Detection via Visual Place
Recognition under Adverse Conditions [29.828201168816243]
We investigate whether a prior map can be leveraged to aid in the detection of dynamic objects in a scene without the need for a 3D map.
We contribute an algorithm which refines an initial set of candidate object detections and produces a refined subset of highly accurate detections using a prior map.
arXiv Detail & Related papers (2023-06-30T10:46:51Z) - TANDEM: Tracking and Dense Mapping in Real-time using Deep Multi-view
Stereo [55.30992853477754]
We present TANDEM, a real-time monocular tracking and dense framework.
For pose estimation, TANDEM performs photometric bundle adjustment based on a sliding window of alignments.
TANDEM shows state-of-the-art real-time 3D reconstruction performance.
arXiv Detail & Related papers (2021-11-14T19:01:02Z) - HM-Net: A Regression Network for Object Center Detection and Tracking on
Wide Area Motion Imagery [1.2249546377051437]
We present our deep neural network-based combined object detection and tracking model, namely, Heat Map Network (HM-Net)
HM-Net is significantly faster than state-of-the-art frame differencing and background subtraction-based methods.
It outperforms state-of-the-art WAMI moving object detection and tracking methods on WPAFB dataset.
arXiv Detail & Related papers (2021-10-19T11:56:30Z) - MFGNet: Dynamic Modality-Aware Filter Generation for RGB-T Tracking [72.65494220685525]
We propose a new dynamic modality-aware filter generation module (named MFGNet) to boost the message communication between visible and thermal data.
We generate dynamic modality-aware filters with two independent networks. The visible and thermal filters will be used to conduct a dynamic convolutional operation on their corresponding input feature maps respectively.
To address issues caused by heavy occlusion, fast motion, and out-of-view, we propose to conduct a joint local and global search by exploiting a new direction-aware target-driven attention mechanism.
arXiv Detail & Related papers (2021-07-22T03:10:51Z) - Salient Objects in Clutter [130.63976772770368]
This paper identifies and addresses a serious design bias of existing salient object detection (SOD) datasets.
This design bias has led to a saturation in performance for state-of-the-art SOD models when evaluated on existing datasets.
We propose a new high-quality dataset and update the previous saliency benchmark.
arXiv Detail & Related papers (2021-05-07T03:49:26Z) - Learning Spatio-Appearance Memory Network for High-Performance Visual
Tracking [79.80401607146987]
Existing object tracking usually learns a bounding-box based template to match visual targets across frames, which cannot accurately learn a pixel-wise representation.
This paper presents a novel segmentation-based tracking architecture, which is equipped with a local-temporal memory network to learn accurate-temporal correspondence.
arXiv Detail & Related papers (2020-09-21T08:12:02Z) - Voxel Map for Visual SLAM [57.07800982410967]
We propose a voxel-map representation to efficiently map points for visual SLAM.
Our method is geometrically guaranteed to fall in the camera field-of-view, and occluded points can be identified and removed to a certain extend.
Experimental results show that our voxel map representation is as efficient as a map with 5s and provides significantly higher localization accuracy (average 46% improvement in RMSE) on the EuRoC dataset.
arXiv Detail & Related papers (2020-03-04T18:39:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.