Multi-View Fusion and Distillation for Subgrade Distresses Detection
based on 3D-GPR
- URL: http://arxiv.org/abs/2308.04779v1
- Date: Wed, 9 Aug 2023 08:06:28 GMT
- Title: Multi-View Fusion and Distillation for Subgrade Distresses Detection
based on 3D-GPR
- Authors: Chunpeng Zhou, Kangjie Ning, Haishuai Wang, Zhi Yu, Sheng Zhou, Jiajun
Bu
- Abstract summary: We introduce a novel methodology for the subgrade distress detection task by leveraging the multi-view information from 3D-GPR data.
We develop a novel textbfMulti-textbfView textbfVusion and textbfDistillation framework, textbfGPR-MVFD, specifically designed to optimally utilize the multi-view GPR dataset.
- Score: 19.49863426864145
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The application of 3D ground-penetrating radar (3D-GPR) for subgrade distress
detection has gained widespread popularity. To enhance the efficiency and
accuracy of detection, pioneering studies have attempted to adopt automatic
detection techniques, particularly deep learning. However, existing works
typically rely on traditional 1D A-scan, 2D B-scan or 3D C-scan data of the
GPR, resulting in either insufficient spatial information or high computational
complexity. To address these challenges, we introduce a novel methodology for
the subgrade distress detection task by leveraging the multi-view information
from 3D-GPR data. Moreover, we construct a real multi-view image dataset
derived from the original 3D-GPR data for the detection task, which provides
richer spatial information compared to A-scan and B-scan data, while reducing
computational complexity compared to C-scan data. Subsequently, we develop a
novel \textbf{M}ulti-\textbf{V}iew \textbf{V}usion and \textbf{D}istillation
framework, \textbf{GPR-MVFD}, specifically designed to optimally utilize the
multi-view GPR dataset. This framework ingeniously incorporates multi-view
distillation and attention-based fusion to facilitate significant feature
extraction for subgrade distresses. In addition, a self-adaptive learning
mechanism is adopted to stabilize the model training and prevent performance
degeneration in each branch. Extensive experiments conducted on this new GPR
benchmark demonstrate the effectiveness and efficiency of our proposed
framework. Our framework outperforms not only the existing GPR baselines, but
also the state-of-the-art methods in the fields of multi-view learning,
multi-modal learning, and knowledge distillation. We will release the
constructed multi-view GPR dataset with expert-annotated labels and the source
codes of the proposed framework.
Related papers
- GSPR: Multimodal Place Recognition Using 3D Gaussian Splatting for Autonomous Driving [9.023864430027333]
multimodal place recognition has gained increasing attention due to their ability to overcome weaknesses of uni sensor systems.
We propose a 3D Gaussian-based multimodal place recognition neural network dubbed GSPR.
arXiv Detail & Related papers (2024-10-01T00:43:45Z) - Cross-Cluster Shifting for Efficient and Effective 3D Object Detection
in Autonomous Driving [69.20604395205248]
We present a new 3D point-based detector model, named Shift-SSD, for precise 3D object detection in autonomous driving.
We introduce an intriguing Cross-Cluster Shifting operation to unleash the representation capacity of the point-based detector.
We conduct extensive experiments on the KITTI, runtime, and nuScenes datasets, and the results demonstrate the state-of-the-art performance of Shift-SSD.
arXiv Detail & Related papers (2024-03-10T10:36:32Z) - VirtualPainting: Addressing Sparsity with Virtual Points and
Distance-Aware Data Augmentation for 3D Object Detection [3.5259183508202976]
We present an innovative approach that involves the generation of virtual LiDAR points using camera images.
We also enhance these virtual points with semantic labels obtained from image-based segmentation networks.
Our approach offers a versatile solution that can be seamlessly integrated into various 3D frameworks and 2D semantic segmentation methods.
arXiv Detail & Related papers (2023-12-26T18:03:05Z) - Domain Generalization of 3D Object Detection by Density-Resampling [14.510085711178217]
Point-cloud-based 3D object detection suffers from performance degradation when encountering data with novel domain gaps.
We propose an SDG method to improve the generalizability of 3D object detection to unseen target domains.
Our work introduces a novel data augmentation method and contributes a new multi-task learning strategy in the methodology.
arXiv Detail & Related papers (2023-11-17T20:01:29Z) - 3DiffTection: 3D Object Detection with Geometry-Aware Diffusion Features [70.50665869806188]
3DiffTection is a state-of-the-art method for 3D object detection from single images.
We fine-tune a diffusion model to perform novel view synthesis conditioned on a single image.
We further train the model on target data with detection supervision.
arXiv Detail & Related papers (2023-11-07T23:46:41Z) - CMR3D: Contextualized Multi-Stage Refinement for 3D Object Detection [57.44434974289945]
We propose Contextualized Multi-Stage Refinement for 3D Object Detection (CMR3D) framework.
Our framework takes a 3D scene as input and strives to explicitly integrate useful contextual information of the scene.
In addition to 3D object detection, we investigate the effectiveness of our framework for the problem of 3D object counting.
arXiv Detail & Related papers (2022-09-13T05:26:09Z) - IAFA: Instance-aware Feature Aggregation for 3D Object Detection from a
Single Image [37.83574424518901]
3D object detection from a single image is an important task in Autonomous Driving.
We propose an instance-aware approach to aggregate useful information for improving the accuracy of 3D object detection.
arXiv Detail & Related papers (2021-03-05T05:47:52Z) - Deep Continuous Fusion for Multi-Sensor 3D Object Detection [103.5060007382646]
We propose a novel 3D object detector that can exploit both LIDAR as well as cameras to perform very accurate localization.
We design an end-to-end learnable architecture that exploits continuous convolutions to fuse image and LIDAR feature maps at different levels of resolution.
arXiv Detail & Related papers (2020-12-20T18:43:41Z) - Improving Point Cloud Semantic Segmentation by Learning 3D Object
Detection [102.62963605429508]
Point cloud semantic segmentation plays an essential role in autonomous driving.
Current 3D semantic segmentation networks focus on convolutional architectures that perform great for well represented classes.
We propose a novel Aware 3D Semantic Detection (DASS) framework that explicitly leverages localization features from an auxiliary 3D object detection task.
arXiv Detail & Related papers (2020-09-22T14:17:40Z) - Spatial-Spectral Residual Network for Hyperspectral Image
Super-Resolution [82.1739023587565]
We propose a novel spectral-spatial residual network for hyperspectral image super-resolution (SSRNet)
Our method can effectively explore spatial-spectral information by using 3D convolution instead of 2D convolution, which enables the network to better extract potential information.
In each unit, we employ spatial and temporal separable 3D convolution to extract spatial and spectral information, which not only reduces unaffordable memory usage and high computational cost, but also makes the network easier to train.
arXiv Detail & Related papers (2020-01-14T03:34:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.