Supervised learning for improving the accuracy of robot-mounted 3D
camera applied to human gait analysis
- URL: http://arxiv.org/abs/2207.01002v1
- Date: Sun, 3 Jul 2022 10:35:18 GMT
- Title: Supervised learning for improving the accuracy of robot-mounted 3D
camera applied to human gait analysis
- Authors: Diego Guffanti, Alberto Brunete, Miguel Hernando, David \'Alvarez,
Javier Rueda, Enrique Navarro
- Abstract summary: The use of 3D cameras for gait analysis has been highly questioned due to the low accuracy they have demonstrated in the past.
The 3D camera was mounted in a mobile robot to obtain a longer walking distance.
This study shows an improvement in detection of kinematic gait signals and gait descriptors by post-processing the raw estimations of the camera.
- Score: 0.31171750528972203
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The use of 3D cameras for gait analysis has been highly questioned due to the
low accuracy they have demonstrated in the past. The objective of the study
presented in this paper is to improve the accuracy of the estimations made by
robot-mounted 3D cameras in human gait analysis by applying a supervised
learning stage. The 3D camera was mounted in a mobile robot to obtain a longer
walking distance. This study shows an improvement in detection of kinematic
gait signals and gait descriptors by post-processing the raw estimations of the
camera using artificial neural networks trained with the data obtained from a
certified Vicon system. To achieve this, 37 healthy participants were recruited
and data of 207 gait sequences were collected using an Orbbec Astra 3D camera.
There are two basic possible approaches for training: using kinematic gait
signals and using gait descriptors. The former seeks to improve the waveforms
of kinematic gait signals by reducing the error and increasing the correlation
with respect to the Vicon system. The second is a more direct approach,
focusing on training the artificial neural networks using gait descriptors
directly. The accuracy of the 3D camera was measured before and after training.
In both training approaches, an improvement was observed. Kinematic gait
signals showed lower errors and higher correlations with respect to the ground
truth. The accuracy of the system to detect gait descriptors also showed a
substantial improvement, mostly for kinematic descriptors rather than
spatio-temporal. When comparing both training approaches, it was not possible
to define which was the absolute best. Therefore, we believe that the selection
of the training approach will depend on the purpose of the study to be
conducted. This study reveals the great potential of 3D cameras and encourages
the research community to continue exploring their use in gait analysis.
Related papers
- Motion Degeneracy in Self-supervised Learning of Elevation Angle
Estimation for 2D Forward-Looking Sonar [4.683630397028384]
This study aims to realize stable self-supervised learning of elevation angle estimation without pretraining using synthetic images.
We first analyze the motion field of 2D forward-looking sonar, which is related to the main supervision signal.
arXiv Detail & Related papers (2023-07-30T08:06:11Z) - Towards Single Camera Human 3D-Kinematics [15.559206592078425]
We propose a novel approach for direct 3D human kinematic estimation D3KE from videos using deep neural networks.
Our experiments demonstrate that the proposed end-to-end training is robust and outperforms 2D and 3D markerless motion capture based kinematic estimation pipelines.
arXiv Detail & Related papers (2023-01-13T08:44:09Z) - An Empirical Study of Pseudo-Labeling for Image-based 3D Object
Detection [72.30883544352918]
We investigate whether pseudo-labels can provide effective supervision for the baseline models under varying settings.
We achieve 20.23 AP for moderate level on the KITTI-3D testing set without bells and whistles, improving the baseline model by 6.03 AP.
We hope this work can provide insights for the image-based 3D detection community under a semi-supervised setting.
arXiv Detail & Related papers (2022-08-15T12:17:46Z) - Neural Scene Representation for Locomotion on Structured Terrain [56.48607865960868]
We propose a learning-based method to reconstruct the local terrain for a mobile robot traversing urban environments.
Using a stream of depth measurements from the onboard cameras and the robot's trajectory, the estimates the topography in the robot's vicinity.
We propose a 3D reconstruction model that faithfully reconstructs the scene, despite the noisy measurements and large amounts of missing data coming from the blind spots of the camera arrangement.
arXiv Detail & Related papers (2022-06-16T10:45:17Z) - Real-Time Human Pose Estimation on a Smart Walker using Convolutional
Neural Networks [4.076099054649463]
We present a novel approach to patient monitoring and data-driven human-in-the-loop control in the context of smart walkers.
It is able to extract a complete and compact body representation in real-time and from inexpensive sensors.
Despite promising results, more data should be collected on users with impairments to assess its performance as a rehabilitation tool in real-world scenarios.
arXiv Detail & Related papers (2021-06-28T14:11:48Z) - AcinoSet: A 3D Pose Estimation Dataset and Baseline Models for Cheetahs
in the Wild [51.35013619649463]
We present an extensive dataset of free-running cheetahs in the wild, called AcinoSet.
The dataset contains 119,490 frames of multi-view synchronized high-speed video footage, camera calibration files and 7,588 human-annotated frames.
The resulting 3D trajectories, human-checked 3D ground truth, and an interactive tool to inspect the data is also provided.
arXiv Detail & Related papers (2021-03-24T15:54:11Z) - Learning the Update Operator for 2D/3D Image Registration [10.720342813316531]
preoperative volume can be overlaid over the 2D images using 2D/3D image registration.
Deep learning-based 2D/3D registration methods have shown promising results by improving computational efficiency and robustness.
We show an improvement of 1.8 times in terms of registration accuracy for the update step prediction compared to learning without the known operator.
arXiv Detail & Related papers (2021-02-04T19:52:59Z) - Synthetic Training for Monocular Human Mesh Recovery [100.38109761268639]
This paper aims to estimate 3D mesh of multiple body parts with large-scale differences from a single RGB image.
The main challenge is lacking training data that have complete 3D annotations of all body parts in 2D images.
We propose a depth-to-scale (D2S) projection to incorporate the depth difference into the projection function to derive per-joint scale variants.
arXiv Detail & Related papers (2020-10-27T03:31:35Z) - Cascaded deep monocular 3D human pose estimation with evolutionary
training data [76.3478675752847]
Deep representation learning has achieved remarkable accuracy for monocular 3D human pose estimation.
This paper proposes a novel data augmentation method that is scalable for massive amount of training data.
Our method synthesizes unseen 3D human skeletons based on a hierarchical human representation and synthesizings inspired by prior knowledge.
arXiv Detail & Related papers (2020-06-14T03:09:52Z) - Lightweight Multi-View 3D Pose Estimation through Camera-Disentangled
Representation [57.11299763566534]
We present a solution to recover 3D pose from multi-view images captured with spatially calibrated cameras.
We exploit 3D geometry to fuse input images into a unified latent representation of pose, which is disentangled from camera view-points.
Our architecture then conditions the learned representation on camera projection operators to produce accurate per-view 2d detections.
arXiv Detail & Related papers (2020-04-05T12:52:29Z) - Lightweight 3D Human Pose Estimation Network Training Using
Teacher-Student Learning [15.321557614896268]
MoVNect is a lightweight deep neural network to capture 3D human pose using a single RGB camera.
We apply the teacher-student learning method based knowledge distillation to 3D human pose estimation.
We implement a 3D avatar application running on mobile in real-time to demonstrate that our network achieves both high accuracy and fast inference time.
arXiv Detail & Related papers (2020-01-15T01:31:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.