SmartPoser: Arm Pose Estimation with a Smartphone and Smartwatch Using UWB and IMU Data
- URL: http://arxiv.org/abs/2509.03451v1
- Date: Wed, 03 Sep 2025 16:16:55 GMT
- Title: SmartPoser: Arm Pose Estimation with a Smartphone and Smartwatch Using UWB and IMU Data
- Authors: Nathan DeVrio, Vimal Mollyn, Chris Harrison,
- Abstract summary: We describe how an off-the-shelf smartphone and smartwatch can work together to accurately estimate arm pose.<n>We take advantage of more recent ultra-wideband (UWB) functionality on these devices to capture absolute distance between the two devices.<n>We quantify the performance of our software-only approach using off-the-shelf devices, showing it can estimate the wrist and elbow joints with a hlmedian positional error of 11.0cm.
- Score: 14.237453119638516
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: The ability to track a user's arm pose could be valuable in a wide range of applications, including fitness, rehabilitation, augmented reality input, life logging, and context-aware assistants. Unfortunately, this capability is not readily available to consumers. Systems either require cameras, which carry privacy issues, or utilize multiple worn IMUs or markers. In this work, we describe how an off-the-shelf smartphone and smartwatch can work together to accurately estimate arm pose. Moving beyond prior work, we take advantage of more recent ultra-wideband (UWB) functionality on these devices to capture absolute distance between the two devices. This measurement is the perfect complement to inertial data, which is relative and suffers from drift. We quantify the performance of our software-only approach using off-the-shelf devices, showing it can estimate the wrist and elbow joints with a \hl{median positional error of 11.0~cm}, without the user having to provide training data.
Related papers
- MOTION: ML-Assisted On-Device Low-Latency Motion Recognition [5.0385144315892925]
We use WeBe Band, a multi-sensor wearable device equipped with a powerful enough MCU to effectively perform gesture recognition entirely on the device.<n>We found that the neural network provided the best balance between accuracy, latency, and memory use.<n>Our results also demonstrate that reliable real-time gesture recognition can be achieved in WeBe Band, with great potential for real-time medical monitoring solutions.
arXiv Detail & Related papers (2025-10-14T01:15:47Z) - TeleOpBench: A Simulator-Centric Benchmark for Dual-Arm Dexterous Teleoperation [50.261933845325636]
We introduce TeleOpBench, a simulator-centric benchmark tailored to bimanual dexterous teleoperation.<n>Within this benchmark we implement four representative teleoperation modalities-(i) MoCap, (ii) VR device, (iii) arm-hand exoskeletons, and (iv) monocular vision tracking.
arXiv Detail & Related papers (2025-05-19T06:08:53Z) - MobilePoser: Real-Time Full-Body Pose Estimation and 3D Human Translation from IMUs in Mobile Consumer Devices [9.50274333425178]
We introduce MobilePoser, a real-time system for full-body pose and global translation estimation.<n>MobilePoser employs a physics-based motion estimation followed by a deep neural network for pose estimation, achieving state-of-the-art accuracy while remaining lightweight.<n>We conclude with a series of applications to illustrate the unique potential of MobilePoser across a variety of fields, such as health and wellness, gaming, and indoor navigation to name a few.
arXiv Detail & Related papers (2025-04-16T21:19:47Z) - Bring Your Rear Cameras for Egocentric 3D Human Pose Estimation [69.68568832269285]
Egocentric 3D human pose estimation has been actively studied using cameras installed in front of a head-mounted device (HMD)<n>It remains unclear if the same holds for full-body tracking due to self-occlusion and limited field-of-view coverage.<n>We propose a new transformer-based method that refines 2D joint heatmap estimation with multi-view information and heatmap uncertainty.
arXiv Detail & Related papers (2025-03-14T17:59:54Z) - MuJo: Multimodal Joint Feature Space Learning for Human Activity Recognition [2.7532797256542403]
Human activity recognition (HAR) is a long-standing problem in artificial intelligence with applications in a broad range of areas.<n>We introduce our comprehensive Fitness Multimodal Activity dataset (FiMAD) to enhance HAR performance across various modalities.<n>We show that FiMAD can increase the performance on real HAR datasets such as MM-Fit, MyoGym, MotionSense, and MHEALTH.
arXiv Detail & Related papers (2024-06-06T08:42:36Z) - Learning Visuotactile Skills with Two Multifingered Hands [80.99370364907278]
We explore learning from human demonstrations using a bimanual system with multifingered hands and visuotactile data.
Our results mark a promising step forward in bimanual multifingered manipulation from visuotactile data.
arXiv Detail & Related papers (2024-04-25T17:59:41Z) - Training-free image style alignment for self-adapting domain shift on
handheld ultrasound devices [54.476120039032594]
We propose the Training-free Image Style Alignment (TISA) framework to align the style of handheld device data to those of standard devices.
TISA can directly infer handheld device images without extra training and is suited for clinical applications.
arXiv Detail & Related papers (2024-02-17T07:15:23Z) - An Examination of Wearable Sensors and Video Data Capture for Human
Exercise Classification [9.674125829493214]
We compare the performance of IMUs to a video-based approach for human exercise classification on two real-world datasets.
We observe that an approach based on a single camera can outperform a single IMU by 10 percentage points on average.
Our work opens up new and more realistic avenues for this application, where a video captured using a readily available smartphone camera, combined with a single sensor, can be used for effective human exercise classification.
arXiv Detail & Related papers (2023-07-10T12:24:04Z) - IMUPoser: Full-Body Pose Estimation using IMUs in Phones, Watches, and
Earbuds [41.8359507387665]
We explore the feasibility of estimating body pose using IMUs already in devices that many users own.
Our pipeline receives whatever subset of IMU data is available, potentially from just a single device, and produces a best-guess pose.
arXiv Detail & Related papers (2023-04-25T02:13:24Z) - SensiX: A Platform for Collaborative Machine Learning on the Edge [69.1412199244903]
We present SensiX, a personal edge platform that stays between sensor data and sensing models.
We demonstrate its efficacy in developing motion and audio-based multi-device sensing systems.
Our evaluation shows that SensiX offers a 7-13% increase in overall accuracy and up to 30% increase across different environment dynamics at the expense of 3mW power overhead.
arXiv Detail & Related papers (2020-12-04T23:06:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.