T-FFTRadNet: Object Detection with Swin Vision Transformers from Raw ADC
Radar Signals
- URL: http://arxiv.org/abs/2303.16940v1
- Date: Wed, 29 Mar 2023 18:04:19 GMT
- Title: T-FFTRadNet: Object Detection with Swin Vision Transformers from Raw ADC
Radar Signals
- Authors: James Giroux, Martin Bouchard, Robert Laganiere
- Abstract summary: Object detection utilizing Frequency Modulated Continous Wave radar is becoming increasingly popular in the field of autonomous systems.
Radar does not possess the same drawbacks seen by other emission-based sensors such as LiDAR, primarily the degradation or loss of return signals due to weather conditions such as rain or snow.
We introduce hierarchical Swin Vision transformers to the field of radar object detection and show their capability to operate on inputs varying in pre-processing, along with different radar configurations.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Object detection utilizing Frequency Modulated Continous Wave radar is
becoming increasingly popular in the field of autonomous systems. Radar does
not possess the same drawbacks seen by other emission-based sensors such as
LiDAR, primarily the degradation or loss of return signals due to weather
conditions such as rain or snow. However, radar does possess traits that make
it unsuitable for standard emission-based deep learning representations such as
point clouds. Radar point clouds tend to be sparse and therefore information
extraction is not efficient. To overcome this, more traditional digital signal
processing pipelines were adapted to form inputs residing directly in the
frequency domain via Fast Fourier Transforms. Commonly, three transformations
were used to form Range-Azimuth-Doppler cubes in which deep learning algorithms
could perform object detection. This too has drawbacks, namely the
pre-processing costs associated with performing multiple Fourier Transforms and
normalization. We explore the possibility of operating on raw radar inputs from
analog to digital converters via the utilization of complex transformation
layers. Moreover, we introduce hierarchical Swin Vision transformers to the
field of radar object detection and show their capability to operate on inputs
varying in pre-processing, along with different radar configurations, i.e.
relatively low and high numbers of transmitters and receivers, while obtaining
on par or better results than the state-of-the-art.
Related papers
- Radar Fields: Frequency-Space Neural Scene Representations for FMCW Radar [62.51065633674272]
We introduce Radar Fields - a neural scene reconstruction method designed for active radar imagers.
Our approach unites an explicit, physics-informed sensor model with an implicit neural geometry and reflectance model to directly synthesize raw radar measurements.
We validate the effectiveness of the method across diverse outdoor scenarios, including urban scenes with dense vehicles and infrastructure.
arXiv Detail & Related papers (2024-05-07T20:44:48Z) - Frequency-Aware Deepfake Detection: Improving Generalizability through
Frequency Space Learning [81.98675881423131]
This research addresses the challenge of developing a universal deepfake detector that can effectively identify unseen deepfake images.
Existing frequency-based paradigms have relied on frequency-level artifacts introduced during the up-sampling in GAN pipelines to detect forgeries.
We introduce a novel frequency-aware approach called FreqNet, centered around frequency domain learning, specifically designed to enhance the generalizability of deepfake detectors.
arXiv Detail & Related papers (2024-03-12T01:28:00Z) - RadarLCD: Learnable Radar-based Loop Closure Detection Pipeline [4.09225917049674]
This research introduces RadarLCD, a novel supervised deep learning pipeline specifically designed for Loop Closure Detection.
RadarLCD makes a significant contribution by leveraging the pre-trained HERO (Hybrid Estimation Radar Odometry) model.
The methodology undergoes evaluation across a variety of FMCW Radar dataset scenes.
arXiv Detail & Related papers (2023-09-13T17:10:23Z) - Echoes Beyond Points: Unleashing the Power of Raw Radar Data in
Multi-modality Fusion [74.84019379368807]
We propose a novel method named EchoFusion to skip the existing radar signal processing pipeline.
Specifically, we first generate the Bird's Eye View (BEV) queries and then take corresponding spectrum features from radar to fuse with other sensors.
arXiv Detail & Related papers (2023-07-31T09:53:50Z) - Semantic Segmentation of Radar Detections using Convolutions on Point
Clouds [59.45414406974091]
We introduce a deep-learning based method to convolve radar detections into point clouds.
We adapt this algorithm to radar-specific properties through distance-dependent clustering and pre-processing of input point clouds.
Our network outperforms state-of-the-art approaches that are based on PointNet++ on the task of semantic segmentation of radar point clouds.
arXiv Detail & Related papers (2023-05-22T07:09:35Z) - RCDPT: Radar-Camera fusion Dense Prediction Transformer [1.5899159309486681]
We propose a novel fusion strategy to integrate radar data into a vision transformer network.
Instead of using readout tokens, radar representations contribute additional depth information to a monocular depth estimation model.
The experiments are conducted on the nuScenes dataset, which includes camera images, lidar, and radar data.
arXiv Detail & Related papers (2022-11-04T13:16:20Z) - mm-Wave Radar Hand Shape Classification Using Deformable Transformers [0.46007387171990594]
A novel, real-time, mm-Wave radar-based static hand shape classification algorithm and implementation are proposed.
The method finds several applications in low cost and privacy sensitive touchless control technology using 60 Ghz radar as the sensor input.
arXiv Detail & Related papers (2022-10-24T09:56:11Z) - Integral Migrating Pre-trained Transformer Encoder-decoders for Visual
Object Detection [78.2325219839805]
imTED improves the state-of-the-art of few-shot object detection by up to 7.6% AP.
Experiments on MS COCO dataset demonstrate that imTED consistently outperforms its counterparts by 2.8%.
arXiv Detail & Related papers (2022-05-19T15:11:20Z) - Unsupervised Domain Adaptation across FMCW Radar Configurations Using
Margin Disparity Discrepancy [17.464353263281907]
In this work, we consider the problem of unsupervised domain adaptation across radar configurations in the context of deep-learning human activity classification.
We focus on the theory-inspired technique of Margin Disparity Discrepancy, which has already been proved successful in the area of computer vision.
Our experiments extend this technique to radar data, achieving a comparable accuracy to fewshot supervised approaches for the same classification problem.
arXiv Detail & Related papers (2022-03-09T09:11:06Z) - Complex-valued Convolutional Neural Networks for Enhanced Radar Signal
Denoising and Interference Mitigation [73.0103413636673]
We propose the use of Complex-Valued Convolutional Neural Networks (CVCNNs) to address the issue of mutual interference between radar sensors.
CVCNNs increase data efficiency, speeds up network training and substantially improves the conservation of phase information during interference removal.
arXiv Detail & Related papers (2021-04-29T10:06:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.