Analyzing Swimming Performance Using Drone Captured Aerial Videos
- URL: http://arxiv.org/abs/2503.12981v1
- Date: Mon, 17 Mar 2025 09:38:44 GMT
- Title: Analyzing Swimming Performance Using Drone Captured Aerial Videos
- Authors: Thu Tran, Kenny Tsu Wei Choo, Shaohui Foong, Hitesh Bhardwaj, Shane Kyi Hla Win, Wei Jun Ang, Kenneth Goh, Rajesh Krishna Balan,
- Abstract summary: This paper presents a novel approach for tracking swimmers using a moving UAV.<n>The proposed system employs a UAV equipped with a high-resolution camera to capture aerial footage of the swimmers.<n>The footage is then processed using computer vision algorithms to extract the swimmers' positions and movements.
- Score: 6.431314461860605
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Monitoring swimmer performance is crucial for improving training and enhancing athletic techniques. Traditional methods for tracking swimmers, such as above-water and underwater cameras, face limitations due to the need for multiple cameras and obstructions from water splashes. This paper presents a novel approach for tracking swimmers using a moving UAV. The proposed system employs a UAV equipped with a high-resolution camera to capture aerial footage of the swimmers. The footage is then processed using computer vision algorithms to extract the swimmers' positions and movements. This approach offers several advantages, including single camera use and comprehensive coverage. The system's accuracy is evaluated with both training and in competition videos. The results demonstrate the system's ability to accurately track swimmers' movements, limb angles, stroke duration and velocity with the maximum error of 0.3 seconds and 0.35~m/s for stroke duration and velocity, respectively.
Related papers
- MegaSaM: Accurate, Fast, and Robust Structure and Motion from Casual Dynamic Videos [104.1338295060383]
We present a system that allows for accurate, fast, and robust estimation of camera parameters and depth maps from casual monocular videos of dynamic scenes.<n>Our system is significantly more accurate and robust at camera pose and depth estimation when compared with prior and concurrent work.
arXiv Detail & Related papers (2024-12-05T18:59:42Z) - AC3D: Analyzing and Improving 3D Camera Control in Video Diffusion Transformers [66.29824750770389]
We analyze camera motion from a first principles perspective, uncovering insights that enable precise 3D camera manipulation.<n>We compound these findings to design the Advanced 3D Camera Control (AC3D) architecture.
arXiv Detail & Related papers (2024-11-27T18:49:13Z) - Investigating Event-Based Cameras for Video Frame Interpolation in Sports [59.755469098797406]
We present a first investigation of event-based Video Frame Interpolation (VFI) models for generating sports slow-motion videos.
Particularly, we design and implement a bi-camera recording setup, including an RGB and an event-based camera to capture sports videos, to temporally align and spatially register both cameras.
Our experimental validation demonstrates that TimeLens, an off-the-shelf event-based VFI model, can effectively generate slow-motion footage for sports videos.
arXiv Detail & Related papers (2024-07-02T15:39:08Z) - Active Human Pose Estimation via an Autonomous UAV Agent [13.188563931419056]
This paper focuses on the task of human pose estimation from videos capturing a person's activity.
To address this, relocating the camera to a new vantage point is necessary to clarify the view.
Our proposed solution comprises three main components: a NeRF-based Drone-View Data Generation Framework, an On-Drone Network for Camera View Error Estimation, and a Combined Planner.
arXiv Detail & Related papers (2024-07-01T21:20:52Z) - SDGE: Stereo Guided Depth Estimation for 360$^\circ$ Camera Sets [65.64958606221069]
Multi-camera systems are often used in autonomous driving to achieve a 360$circ$ perception.
These 360$circ$ camera sets often have limited or low-quality overlap regions, making multi-view stereo methods infeasible for the entire image.
We propose the Stereo Guided Depth Estimation (SGDE) method, which enhances depth estimation of the full image by explicitly utilizing multi-view stereo results on the overlap.
arXiv Detail & Related papers (2024-02-19T02:41:37Z) - SwimXYZ: A large-scale dataset of synthetic swimming motions and videos [0.0]
We introduce SwimXYZ, a synthetic dataset of swimming motions and videos.
SwimXYZ contains 3.4 million frames annotated with ground truth 2D and 3D joints, as well as 240 sequences of swimming motions in the SMPL parameters format.
In addition to making this dataset publicly available, we present use cases for SwimXYZ in swimming stroke clustering and 2D pose estimation.
arXiv Detail & Related papers (2023-10-06T16:33:24Z) - AZTR: Aerial Video Action Recognition with Auto Zoom and Temporal
Reasoning [63.628195002143734]
We propose a novel approach for aerial video action recognition.
Our method is designed for videos captured using UAVs and can run on edge or mobile devices.
We present a learning-based approach that uses customized auto zoom to automatically identify the human target and scale it appropriately.
arXiv Detail & Related papers (2023-03-02T21:24:19Z) - Shakes on a Plane: Unsupervised Depth Estimation from Unstabilized
Photography [54.36608424943729]
We show that in a ''long-burst'', forty-two 12-megapixel RAW frames captured in a two-second sequence, there is enough parallax information from natural hand tremor alone to recover high-quality scene depth.
We devise a test-time optimization approach that fits a neural RGB-D representation to long-burst data and simultaneously estimates scene depth and camera motion.
arXiv Detail & Related papers (2022-12-22T18:54:34Z) - Spatiotemporal Motion Synchronization for Snowboard Big Air [6.054558868204333]
We propose a conventional but plausible solution using existing image processing techniques for snowboard big air training.
We conducted interviews with expert snowboarders who stated that thetemporally aligned videos enabled them to precisely identify slight differences in body movements.
arXiv Detail & Related papers (2021-12-20T23:30:33Z) - Swimmer Stroke Rate Estimation From Overhead Race Video [26.24508656138528]
We propose a system for automatically determining swimmer stroke rates from overhead race video (ORV)
General ORV is defined as any footage of swimmers in competition, taken for the purposes of viewing or analysis.
We detail how to create a system that will automatically collect swimmer stroke rates in any competition, given the video of the competition of interest.
arXiv Detail & Related papers (2021-04-25T04:20:38Z) - Do You See What I See? Coordinating Multiple Aerial Cameras for Robot
Cinematography [9.870369982132678]
We develop a real-time multi-UAV coordination system that is capable of recording dynamic targets while maximizing shot diversity and avoiding collisions.
We show that our coordination scheme has low computational cost and takes only 1.17 ms on average to plan for a team of 3 UAVs over a 10 s time horizon.
arXiv Detail & Related papers (2020-11-10T22:43:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.