Fine-grained Hand Gesture Recognition in Multi-viewpoint Hand Hygiene
- URL: http://arxiv.org/abs/2109.02917v1
- Date: Tue, 7 Sep 2021 08:14:15 GMT
- Title: Fine-grained Hand Gesture Recognition in Multi-viewpoint Hand Hygiene
- Authors: Huy Q.Vo, Tuong Do, Vi C.Pham, Duy Nguyen, An T.Duong and Quang D.Tran
- Abstract summary: This paper contributes a new high-quality dataset for hand gesture recognition in hand hygiene systems, named "MFH"
To address the aforementioned issues, the MFH dataset is proposed to contain a total of 731147 samples obtained by different camera views in 6 non-overlapping locations.
- Score: 3.588453140011797
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper contributes a new high-quality dataset for hand gesture
recognition in hand hygiene systems, named "MFH". Generally, current datasets
are not focused on: (i) fine-grained actions; and (ii) data mismatch between
different viewpoints, which are available under realistic settings. To address
the aforementioned issues, the MFH dataset is proposed to contain a total of
731147 samples obtained by different camera views in 6 non-overlapping
locations. Additionally, each sample belongs to one of seven steps introduced
by the World Health Organization (WHO). As a minor contribution, inspired by
advances in fine-grained image recognition and distribution adaptation, this
paper recommends using the self-supervised learning method to handle these
preceding problems. The extensive experiments on the benchmarking MFH dataset
show that the introduced method yields competitive performance in both the
Accuracy and the Macro F1-score. The code and the MFH dataset are available at
https://github.com/willogy-team/hand-gesture-recognition-smc2021.
Related papers
- Beyond Confusion: A Fine-grained Dialectical Examination of Human Activity Recognition Benchmark Datasets [4.249657064343807]
The research of machine learning algorithms for human activity recognition (HAR) has made significant progress with publicly available datasets.
Recent models like transformers have been applied to HAR datasets with limited success from the benchmark metrics.
This paper aims to address these open questions by conducting a fine-grained inspection of six popular HAR benchmark datasets.
arXiv Detail & Related papers (2024-12-12T07:53:17Z) - Benchmarking Federated Learning for Semantic Datasets: Federated Scene Graph Generation [3.499870393443268]
Federated learning (FL) has recently garnered attention as a data-decentralized training framework.
We propose a benchmark process to establish an FL benchmark with controllable semantic heterogeneity across clients.
As a proof of concept, we first construct a federated PSG benchmark, demonstrating the efficacy of the existing PSG methods in an FL setting.
arXiv Detail & Related papers (2024-12-11T08:10:46Z) - SMILE-UHURA Challenge -- Small Vessel Segmentation at Mesoscopic Scale from Ultra-High Resolution 7T Magnetic Resonance Angiograms [60.35639972035727]
The lack of publicly available annotated datasets has impeded the development of robust, machine learning-driven segmentation algorithms.
The SMILE-UHURA challenge addresses the gap in publicly available annotated datasets by providing an annotated dataset of Time-of-Flight angiography acquired with 7T MRI.
Dice scores reached up to 0.838 $pm$ 0.066 and 0.716 $pm$ 0.125 on the respective datasets, with an average performance of up to 0.804 $pm$ 0.15.
arXiv Detail & Related papers (2024-11-14T17:06:00Z) - Multi-fingered Robotic Hand Grasping in Cluttered Environments through Hand-object Contact Semantic Mapping [14.674925349389179]
We develop a method for generating multi-fingered hand grasp samples in cluttered settings through contact semantic map.
We also propose the multi-modal multi-fingered grasping dataset generation method.
arXiv Detail & Related papers (2024-04-12T23:11:36Z) - Domain Adaptive Synapse Detection with Weak Point Annotations [63.97144211520869]
We present AdaSyn, a framework for domain adaptive synapse detection with weak point annotations.
In the WASPSYN challenge at I SBI 2023, our method ranks the 1st place.
arXiv Detail & Related papers (2023-08-31T05:05:53Z) - Multi-Granularity Hand Action Detection [58.88274905101276]
FHA-Kitchens dataset comprises 2,377 video clips and 30,047 frames, annotated with approximately 200k bounding boxes and 880 action categories.
This dataset comprises 2,377 video clips and 30,047 frames, annotated with approximately 200k bounding boxes and 880 action categories.
We propose MG-HAD, an End-to-End Multi-Granularity Hand Action Detection method.
arXiv Detail & Related papers (2023-06-19T11:21:59Z) - Exploring the Boundaries of Semi-Supervised Facial Expression Recognition using In-Distribution, Out-of-Distribution, and Unconstrained Data [23.4909421082857]
We present a study on 11 of the most recent semi-supervised methods, in the context of facial expression recognition (FER)
Our investigation covers semi-supervised learning from in-distribution, out-of-distribution, unconstrained, and very small unlabelled data.
With an equal number of labelled samples, semi-supervised learning delivers a considerable improvement over supervised learning.
arXiv Detail & Related papers (2023-06-02T01:40:08Z) - The Second Monocular Depth Estimation Challenge [93.1678025923996]
The second edition of the Monocular Depth Estimation Challenge (MDEC) was open to methods using any form of supervision.
The challenge was based around the SYNS-Patches dataset, which features a wide diversity of environments with high-quality dense ground-truth.
The top supervised submission improved relative F-Score by 27.62%, while the top self-supervised improved it by 16.61%.
arXiv Detail & Related papers (2023-04-14T11:10:07Z) - Cluster-level pseudo-labelling for source-free cross-domain facial
expression recognition [94.56304526014875]
We propose the first Source-Free Unsupervised Domain Adaptation (SFUDA) method for Facial Expression Recognition (FER)
Our method exploits self-supervised pretraining to learn good feature representations from the target data.
We validate the effectiveness of our method in four adaptation setups, proving that it consistently outperforms existing SFUDA methods when applied to FER.
arXiv Detail & Related papers (2022-10-11T08:24:50Z) - Omni-supervised Facial Expression Recognition via Distilled Data [120.11782405714234]
We propose omni-supervised learning to exploit reliable samples in a large amount of unlabeled data for network training.
We experimentally verify that the new dataset can significantly improve the ability of the learned FER model.
To tackle this, we propose to apply a dataset distillation strategy to compress the created dataset into several informative class-wise images.
arXiv Detail & Related papers (2020-05-18T09:36:51Z) - ROAM: Random Layer Mixup for Semi-Supervised Learning in Medical Imaging [43.26668942258135]
Medical image segmentation is one of the major challenges addressed by machine learning methods.
We propose ROAM, a RandOm lAyer Mixup, which generates more data points that have never seen before.
ROAM achieves state-of-the-art (SOTA) results in fully supervised (89.5%) and semi-supervised (87.0%) settings with a relative improvement of up to 2.40% and 16.50%, respectively for the whole-brain segmentation.
arXiv Detail & Related papers (2020-03-20T18:07:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.