RTNH+: Enhanced 4D Radar Object Detection Network using Combined
CFAR-based Two-level Preprocessing and Vertical Encoding
- URL: http://arxiv.org/abs/2310.17659v1
- Date: Thu, 19 Oct 2023 06:45:19 GMT
- Title: RTNH+: Enhanced 4D Radar Object Detection Network using Combined
CFAR-based Two-level Preprocessing and Vertical Encoding
- Authors: Seung-Hyun Kong, Dong-Hee Paek, Sangjae Cho
- Abstract summary: RTNH+ is an enhanced version of RTNH, a 4D Radar object detection network.
We show that RTNH+ achieves significant performance improvement of 10.14% in $AP_3DIoU=0.3$ and 16.12% in $AP_3DIoU=0.5$ over RTNH.
- Score: 8.017543518311196
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Four-dimensional (4D) Radar is a useful sensor for 3D object detection and
the relative radial speed estimation of surrounding objects under various
weather conditions. However, since Radar measurements are corrupted with
invalid components such as noise, interference, and clutter, it is necessary to
employ a preprocessing algorithm before the 3D object detection with neural
networks. In this paper, we propose RTNH+ that is an enhanced version of RTNH,
a 4D Radar object detection network, by two novel algorithms. The first
algorithm is the combined constant false alarm rate (CFAR)-based two-level
preprocessing (CCTP) algorithm that generates two filtered measurements of
different characteristics using the same 4D Radar measurements, which can
enrich the information of the input to the 4D Radar object detection network.
The second is the vertical encoding (VE) algorithm that effectively encodes
vertical features of the road objects from the CCTP outputs. We provide details
of the RTNH+, and demonstrate that RTNH+ achieves significant performance
improvement of 10.14\% in ${{AP}_{3D}^{IoU=0.3}}$ and 16.12\% in
${{AP}_{3D}^{IoU=0.5}}$ over RTNH.
Related papers
- Multistream Network for LiDAR and Camera-based 3D Object Detection in Outdoor Scenes [59.78696921486972]
Fusion of LiDAR and RGB data has the potential to enhance outdoor 3D object detection accuracy.<n>We propose a MultiStream Detection (MuStD) network, that meticulously extracts task-relevant information from both data modalities.
arXiv Detail & Related papers (2025-07-25T14:20:16Z) - SpikingRTNH: Spiking Neural Network for 4D Radar Object Detection [6.636342419996716]
SpikingRTNH is the first spiking neural network (SNN) for 3D object detection using 4D Radar data.
We introduce biological top-down inference (BTI) which processes point clouds sequentially from higher to lower densities.
Results establish the viability of SNNs for energy-efficient 4D Radar-based object detection in autonomous driving systems.
arXiv Detail & Related papers (2025-01-31T07:33:30Z) - GET-UP: GEomeTric-aware Depth Estimation with Radar Points UPsampling [7.90238039959534]
Existing algorithms process radar data by projecting 3D points onto the image plane for pixel-level feature extraction.
We propose GET-UP, leveraging attention-enhanced Graph Neural Networks (GNN) to exchange and aggregate both 2D and 3D information from radar data.
We benchmark our proposed GET-UP on the nuScenes dataset, achieving state-of-the-art performance with a 15.3% and 14.7% improvement in MAE and RMSE over the previously best-performing model.
arXiv Detail & Related papers (2024-09-02T14:15:09Z) - Efficient 4D Radar Data Auto-labeling Method using LiDAR-based Object Detection Network [5.405156980077946]
Existing 4D radar datasets lack sufficient sensor data and labels.
To address these issues, we propose the auto-labeling method of 4D radar tensor (4DRT) in the K-Radar dataset.
arXiv Detail & Related papers (2024-05-13T04:28:06Z) - 3DiffTection: 3D Object Detection with Geometry-Aware Diffusion Features [70.50665869806188]
3DiffTection is a state-of-the-art method for 3D object detection from single images.
We fine-tune a diffusion model to perform novel view synthesis conditioned on a single image.
We further train the model on target data with detection supervision.
arXiv Detail & Related papers (2023-11-07T23:46:41Z) - 4DRVO-Net: Deep 4D Radar-Visual Odometry Using Multi-Modal and
Multi-Scale Adaptive Fusion [2.911052912709637]
Four-dimensional (4D) radar--visual odometry (4DRVO) integrates complementary information from 4D radar and cameras.
4DRVO may exhibit significant tracking errors owing to sparsity of 4D radar point clouds.
We present 4DRVO-Net, which is a method for 4D radar--visual odometry.
arXiv Detail & Related papers (2023-08-12T14:00:09Z) - V-DETR: DETR with Vertex Relative Position Encoding for 3D Object
Detection [73.37781484123536]
We introduce a highly performant 3D object detector for point clouds using the DETR framework.
To address the limitation, we introduce a novel 3D Relative Position (3DV-RPE) method.
We show exceptional results on the challenging ScanNetV2 benchmark.
arXiv Detail & Related papers (2023-08-08T17:14:14Z) - SMURF: Spatial Multi-Representation Fusion for 3D Object Detection with
4D Imaging Radar [12.842457981088378]
This paper introduces spatial multi-representation fusion (SMURF), a novel approach to 3D object detection using a single 4D imaging radar.
SMURF mitigates measurement inaccuracy caused by limited angular resolution and multi-path propagation of radar signals.
Experimental evaluations on View-of-Delft (VoD) and TJ4DRadSet datasets demonstrate the effectiveness and generalization ability of SMURF.
arXiv Detail & Related papers (2023-07-20T11:33:46Z) - Minkowski Tracker: A Sparse Spatio-Temporal R-CNN for Joint Object
Detection and Tracking [53.64390261936975]
We present Minkowski Tracker, a sparse-temporal R-CNN that jointly solves object detection and tracking problems.
Inspired by region-based CNN (R-CNN), we propose to track motion as a second stage of the object detector R-CNN.
We show in large-scale experiments that the overall performance gain of our method is due to four factors.
arXiv Detail & Related papers (2022-08-22T04:47:40Z) - PLUME: Efficient 3D Object Detection from Stereo Images [95.31278688164646]
Existing methods tackle the problem in two steps: first depth estimation is performed, a pseudo LiDAR point cloud representation is computed from the depth estimates, and then object detection is performed in 3D space.
We propose a model that unifies these two tasks in the same metric space.
Our approach achieves state-of-the-art performance on the challenging KITTI benchmark, with significantly reduced inference time compared with existing methods.
arXiv Detail & Related papers (2021-01-17T05:11:38Z) - Dynamic Edge Weights in Graph Neural Networks for 3D Object Detection [0.0]
We propose an attention based feature aggregation technique in graph neural network (GNN) for detecting objects in LiDAR scan.
In each layer of the GNN, apart from the linear transformation which maps the per node input features to the corresponding higher level features, a per node masked attention is also performed.
The experiments on KITTI dataset show that our method yields comparable results for 3D object detection.
arXiv Detail & Related papers (2020-09-17T12:56:17Z) - End-to-End Pseudo-LiDAR for Image-Based 3D Object Detection [62.34374949726333]
Pseudo-LiDAR (PL) has led to a drastic reduction in the accuracy gap between methods based on LiDAR sensors and those based on cheap stereo cameras.
PL combines state-of-the-art deep neural networks for 3D depth estimation with those for 3D object detection by converting 2D depth map outputs to 3D point cloud inputs.
We introduce a new framework based on differentiable Change of Representation (CoR) modules that allow the entire PL pipeline to be trained end-to-end.
arXiv Detail & Related papers (2020-04-07T02:18:38Z) - DSGN: Deep Stereo Geometry Network for 3D Object Detection [79.16397166985706]
There is a large performance gap between image-based and LiDAR-based 3D object detectors.
Our method, called Deep Stereo Geometry Network (DSGN), significantly reduces this gap.
For the first time, we provide a simple and effective one-stage stereo-based 3D detection pipeline.
arXiv Detail & Related papers (2020-01-10T11:44:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.