UltraGlove: Hand Pose Estimation with Mems-Ultrasonic Sensors
- URL: http://arxiv.org/abs/2306.12652v2
- Date: Thu, 14 Sep 2023 22:56:01 GMT
- Title: UltraGlove: Hand Pose Estimation with Mems-Ultrasonic Sensors
- Authors: Qiang Zhang, Yuanqiao Lin, Yubin Lin, Szymon Rusinkiewicz
- Abstract summary: We propose a novel and low-cost hand-tracking glove that utilizes several MEMS-ultrasonic sensors attached to the fingers.
Our experimental results demonstrate that this approach is both accurate, size-agnostic, and robust to external interference.
- Score: 14.257535961674021
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Hand tracking is an important aspect of human-computer interaction and has a
wide range of applications in extended reality devices. However, current hand
motion capture methods suffer from various limitations. For instance,
visual-based hand pose estimation is susceptible to self-occlusion and changes
in lighting conditions, while IMU-based tracking gloves experience significant
drift and are not resistant to external magnetic field interference. To address
these issues, we propose a novel and low-cost hand-tracking glove that utilizes
several MEMS-ultrasonic sensors attached to the fingers, to measure the
distance matrix among the sensors. Our lightweight deep network then
reconstructs the hand pose from the distance matrix. Our experimental results
demonstrate that this approach is both accurate, size-agnostic, and robust to
external interference. We also show the design logic for the sensor selection,
sensor configurations, circuit diagram, as well as model architecture.
Related papers
- Capturing complex hand movements and object interactions using machine learning-powered stretchable smart textile gloves [9.838013581109681]
Real-time tracking of dexterous hand movements has numerous applications in human-computer interaction, metaverse, robotics, and tele-health.
Here, we report accurate and dynamic tracking of articulated hand and finger movements using stretchable, washable smart gloves with embedded helical sensor yarns and inertial measurement units.
The sensor yarns have a high dynamic range, responding to low 0.005 % to high 155 % strains, and show stability during extensive use and washing cycles.
arXiv Detail & Related papers (2024-10-03T05:32:16Z) - Multimodal Active Measurement for Human Mesh Recovery in Close Proximity [13.265259738826302]
In physical human-robot interactions, a robot needs to estimate the accurate body pose of a target person.
In these pHRI scenarios, the robot cannot fully observe the target person's body with equipped cameras because the target person must be close to the robot for physical interaction.
We propose an active measurement and sensor fusion framework of the equipped cameras with touch and ranging sensors such as 2D LiDAR.
arXiv Detail & Related papers (2023-10-12T08:17:57Z) - DiffusionPoser: Real-time Human Motion Reconstruction From Arbitrary Sparse Sensors Using Autoregressive Diffusion [10.439802168557513]
Motion capture from a limited number of body-worn sensors has important applications in health, human performance, and entertainment.
Recent work has focused on accurately reconstructing whole-body motion from a specific sensor configuration using six IMUs.
We propose a single diffusion model, DiffusionPoser, which reconstructs human motion in real-time from an arbitrary combination of sensors.
arXiv Detail & Related papers (2023-08-31T12:36:50Z) - Multi-Modal Neural Radiance Field for Monocular Dense SLAM with a
Light-Weight ToF Sensor [58.305341034419136]
We present the first dense SLAM system with a monocular camera and a light-weight ToF sensor.
We propose a multi-modal implicit scene representation that supports rendering both the signals from the RGB camera and light-weight ToF sensor.
Experiments demonstrate that our system well exploits the signals of light-weight ToF sensors and achieves competitive results.
arXiv Detail & Related papers (2023-08-28T07:56:13Z) - Collision-aware In-hand 6D Object Pose Estimation using Multiple
Vision-based Tactile Sensors [4.886250215151643]
We reason on the possible spatial configurations of the sensors along the object surface.
We use selected sensors configurations to optimize over the space of 6D poses.
We rank the obtained poses by penalizing those that are in collision with the sensors.
arXiv Detail & Related papers (2023-01-31T14:35:26Z) - Reconfigurable Data Glove for Reconstructing Physical and Virtual Grasps [100.72245315180433]
We present a reconfigurable data glove design to capture different modes of human hand-object interactions.
The glove operates in three modes for various downstream tasks with distinct features.
We evaluate the system's three modes by (i) recording hand gestures and associated forces, (ii) improving manipulation fluency in VR, and (iii) producing realistic simulation effects of various tool uses.
arXiv Detail & Related papers (2023-01-14T05:35:50Z) - DensePose From WiFi [86.61881052177228]
We develop a deep neural network that maps the phase and amplitude of WiFi signals to UV coordinates within 24 human regions.
Our model can estimate the dense pose of multiple subjects, with comparable performance to image-based approaches.
arXiv Detail & Related papers (2022-12-31T16:48:43Z) - Drone Detection and Tracking in Real-Time by Fusion of Different Sensing
Modalities [66.4525391417921]
We design and evaluate a multi-sensor drone detection system.
Our solution integrates a fish-eye camera as well to monitor a wider part of the sky and steer the other cameras towards objects of interest.
The thermal camera is shown to be a feasible solution as good as the video camera, even if the camera employed here has a lower resolution.
arXiv Detail & Related papers (2022-07-05T10:00:58Z) - Learning Online Multi-Sensor Depth Fusion [100.84519175539378]
SenFuNet is a depth fusion approach that learns sensor-specific noise and outlier statistics.
We conduct experiments with various sensor combinations on the real-world CoRBS and Scene3D datasets.
arXiv Detail & Related papers (2022-04-07T10:45:32Z) - Monocular Depth Estimation for Soft Visuotactile Sensors [24.319343057803973]
We investigate the application of state-of-the-art monocular depth estimation to infer dense internal (tactile) depth maps directly from an internal single small IR imaging sensor.
We show that deep networks typically used for long-range depth estimation (1-100m) can be effectively trained for precise predictions at a much shorter range (1-100mm) inside a mostly textureless deformable fluid-filled sensor.
We propose a simple supervised learning process to train an object-agnostic network requiring less than 10 random poses in contact for less than 10 seconds for a small set of diverse objects.
arXiv Detail & Related papers (2021-01-05T17:51:11Z) - OmniTact: A Multi-Directional High Resolution Touch Sensor [109.28703530853542]
Existing tactile sensors are either flat, have small sensitive fields or only provide low-resolution signals.
We introduce OmniTact, a multi-directional high-resolution tactile sensor.
We evaluate the capabilities of OmniTact on a challenging robotic control task.
arXiv Detail & Related papers (2020-03-16T01:31:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.