HuPR: A Benchmark for Human Pose Estimation Using Millimeter Wave Radar
- URL: http://arxiv.org/abs/2210.12564v1
- Date: Sat, 22 Oct 2022 22:28:40 GMT
- Title: HuPR: A Benchmark for Human Pose Estimation Using Millimeter Wave Radar
- Authors: Shih-Po Lee, Niraj Prakash Kini, Wen-Hsiao Peng, Ching-Wen Ma,
Jenq-Neng Hwang
- Abstract summary: This paper introduces a novel human pose estimation benchmark, Human Pose with Millimeter Wave Radar (HuPR)
This dataset is created using cross-calibrated mmWave radar sensors and a monocular RGB camera for cross-modality training of radar-based human pose estimation.
- Score: 30.51398364813315
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: This paper introduces a novel human pose estimation benchmark, Human Pose
with Millimeter Wave Radar (HuPR), that includes synchronized vision and radio
signal components. This dataset is created using cross-calibrated mmWave radar
sensors and a monocular RGB camera for cross-modality training of radar-based
human pose estimation. There are two advantages of using mmWave radar to
perform human pose estimation. First, it is robust to dark and low-light
conditions. Second, it is not visually perceivable by humans and thus, can be
widely applied to applications with privacy concerns, e.g., surveillance
systems in patient rooms. In addition to the benchmark, we propose a
cross-modality training framework that leverages the ground-truth 2D keypoints
representing human body joints for training, which are systematically generated
from the pre-trained 2D pose estimation network based on a monocular camera
input image, avoiding laborious manual label annotation efforts. The framework
consists of a new radar pre-processing method that better extracts the velocity
information from radar data, Cross- and Self-Attention Module (CSAM), to fuse
multi-scale radar features, and Pose Refinement Graph Convolutional Networks
(PRGCN), to refine the predicted keypoint confidence heatmaps. Our intensive
experiments on the HuPR benchmark show that the proposed scheme achieves better
human pose estimation performance with only radar data, as compared to
traditional pre-processing solutions and previous radio-frequency-based
methods.
Related papers
- CaFNet: A Confidence-Driven Framework for Radar Camera Depth Estimation [6.9404362058736995]
This paper introduces a two-stage, end-to-end trainable Confidence-aware Fusion Net (CaFNet) for dense depth estimation.
The first stage addresses radar-specific challenges, such as ambiguous elevation and noisy measurements.
For the final depth estimation, we innovate a confidence-aware gated fusion mechanism to integrate radar and image features effectively.
arXiv Detail & Related papers (2024-06-30T13:39:29Z) - ProbRadarM3F: mmWave Radar based Human Skeletal Pose Estimation with Probability Map Guided Multi-Format Feature Fusion [14.83158440666821]
This paper introduces a probability map guided multi-format feature fusion model, ProbRadarM3F.
ProbRadarM3F fuses the traditional heatmap features and the positional features, then effectively achieves the estimation of 14 keypoints of the human body.
arXiv Detail & Related papers (2024-05-08T15:54:57Z) - Radar Fields: Frequency-Space Neural Scene Representations for FMCW Radar [62.51065633674272]
We introduce Radar Fields - a neural scene reconstruction method designed for active radar imagers.
Our approach unites an explicit, physics-informed sensor model with an implicit neural geometry and reflectance model to directly synthesize raw radar measurements.
We validate the effectiveness of the method across diverse outdoor scenarios, including urban scenes with dense vehicles and infrastructure.
arXiv Detail & Related papers (2024-05-07T20:44:48Z) - Diffusion Model is a Good Pose Estimator from 3D RF-Vision [32.72703340013302]
Human pose estimation (HPE) from Radio Frequency vision (RF-vision) performs human sensing using RF signals.
mmWave radar has emerged as a promising RF-vision sensor, providing radar point clouds by processing RF signals.
This work proposes mmDiff, a novel diffusion-based pose estimator tailored for noisy radar data.
arXiv Detail & Related papers (2024-03-24T15:39:52Z) - Echoes Beyond Points: Unleashing the Power of Raw Radar Data in
Multi-modality Fusion [74.84019379368807]
We propose a novel method named EchoFusion to skip the existing radar signal processing pipeline.
Specifically, we first generate the Bird's Eye View (BEV) queries and then take corresponding spectrum features from radar to fuse with other sensors.
arXiv Detail & Related papers (2023-07-31T09:53:50Z) - Semantic Segmentation of Radar Detections using Convolutions on Point
Clouds [59.45414406974091]
We introduce a deep-learning based method to convolve radar detections into point clouds.
We adapt this algorithm to radar-specific properties through distance-dependent clustering and pre-processing of input point clouds.
Our network outperforms state-of-the-art approaches that are based on PointNet++ on the task of semantic segmentation of radar point clouds.
arXiv Detail & Related papers (2023-05-22T07:09:35Z) - DensePose From WiFi [86.61881052177228]
We develop a deep neural network that maps the phase and amplitude of WiFi signals to UV coordinates within 24 human regions.
Our model can estimate the dense pose of multiple subjects, with comparable performance to image-based approaches.
arXiv Detail & Related papers (2022-12-31T16:48:43Z) - Cross Vision-RF Gait Re-identification with Low-cost RGB-D Cameras and
mmWave Radars [15.662787088335618]
This work studies the problem of cross-modal human re-identification (ReID)
We propose the first-of-its-kind vision-RF system for cross-modal multi-person ReID at the same time.
Our proposed system is able to achieve 92.5% top-1 accuracy and 97.5% top-5 accuracy out of 56 volunteers.
arXiv Detail & Related papers (2022-07-16T10:34:25Z) - RadarLoc: Learning to Relocalize in FMCW Radar [36.68888832365474]
We propose a novel end-to-end neural network with self-attention, termed RadarLoc, which is able to estimate 6-DoF global poses directly.
We validate our approach on the recently released challenging outdoor dataset Oxford Radar RobotCar.
arXiv Detail & Related papers (2021-03-22T03:22:37Z) - Depth Estimation from Monocular Images and Sparse Radar Data [93.70524512061318]
In this paper, we explore the possibility of achieving a more accurate depth estimation by fusing monocular images and Radar points using a deep neural network.
We find that the noise existing in Radar measurements is one of the main key reasons that prevents one from applying the existing fusion methods.
The experiments are conducted on the nuScenes dataset, which is one of the first datasets which features Camera, Radar, and LiDAR recordings in diverse scenes and weather conditions.
arXiv Detail & Related papers (2020-09-30T19:01:33Z) - RadarNet: Exploiting Radar for Robust Perception of Dynamic Objects [73.80316195652493]
We tackle the problem of exploiting Radar for perception in the context of self-driving cars.
We propose a new solution that exploits both LiDAR and Radar sensors for perception.
Our approach, dubbed RadarNet, features a voxel-based early fusion and an attention-based late fusion.
arXiv Detail & Related papers (2020-07-28T17:15:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.