Pushing the Envelope for Depth-Based Semi-Supervised 3D Hand Pose
Estimation with Consistency Training
- URL: http://arxiv.org/abs/2303.15147v1
- Date: Mon, 27 Mar 2023 12:32:49 GMT
- Title: Pushing the Envelope for Depth-Based Semi-Supervised 3D Hand Pose
Estimation with Consistency Training
- Authors: Mohammad Rezaei, Farnaz Farahanipad, Alex Dillhoff, Vassilis Athitsos
- Abstract summary: We propose a semi-supervised method to significantly reduce the dependence on labeled training data.
The proposed method consists of two identical networks trained jointly: a teacher network and a student network.
Experiments demonstrate that the proposed method outperforms the state-of-the-art semi-supervised methods by large margins.
- Score: 2.6954666679827137
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite the significant progress that depth-based 3D hand pose estimation
methods have made in recent years, they still require a large amount of labeled
training data to achieve high accuracy. However, collecting such data is both
costly and time-consuming. To tackle this issue, we propose a semi-supervised
method to significantly reduce the dependence on labeled training data. The
proposed method consists of two identical networks trained jointly: a teacher
network and a student network. The teacher network is trained using both the
available labeled and unlabeled samples. It leverages the unlabeled samples via
a loss formulation that encourages estimation equivariance under a set of
affine transformations. The student network is trained using the unlabeled
samples with their pseudo-labels provided by the teacher network. For inference
at test time, only the student network is used. Extensive experiments
demonstrate that the proposed method outperforms the state-of-the-art
semi-supervised methods by large margins.
Related papers
- TrajSSL: Trajectory-Enhanced Semi-Supervised 3D Object Detection [59.498894868956306]
Pseudo-labeling approaches to semi-supervised learning adopt a teacher-student framework.
We leverage pre-trained motion-forecasting models to generate object trajectories on pseudo-labeled data.
Our approach improves pseudo-label quality in two distinct manners.
arXiv Detail & Related papers (2024-09-17T05:35:00Z) - Doubly Robust Self-Training [46.168395767948965]
We introduce doubly robust self-training, a novel semi-supervised algorithm.
We demonstrate the superiority of the doubly robust loss over the standard self-training baseline.
arXiv Detail & Related papers (2023-06-01T00:57:16Z) - Analysis of Semi-Supervised Methods for Facial Expression Recognition [19.442685015494316]
Training deep neural networks for image recognition often requires large-scale human annotated data.
Semi-supervised methods have been proposed to reduce the reliance of deep neural solutions on labeled data.
Our study shows that when training existing semi-supervised methods on as little as 250 labeled samples per class can yield comparable performances to that of fully-supervised methods trained on the full labeled datasets.
arXiv Detail & Related papers (2022-07-31T23:58:35Z) - Learning from Data with Noisy Labels Using Temporal Self-Ensemble [11.245833546360386]
Deep neural networks (DNNs) have an enormous capacity to memorize noisy labels.
Current state-of-the-art methods present a co-training scheme that trains dual networks using samples associated with small losses.
We propose a simple yet effective robust training scheme that operates by training only a single network.
arXiv Detail & Related papers (2022-07-21T08:16:31Z) - Semi-Supervised Training to Improve Player and Ball Detection in Soccer [11.376125584750548]
We present a novel semi-supervised method to train a network based on a labeled image dataset by leveraging a large unlabeled dataset of soccer broadcast videos.
We show that including unlabeled data in the training process allows to substantially improve the performances of the detection network trained only on the labeled data.
arXiv Detail & Related papers (2022-04-14T10:20:56Z) - Active Learning for Deep Visual Tracking [51.5063680734122]
Convolutional neural networks (CNNs) have been successfully applied to the single target tracking task in recent years.
In this paper, we propose an active learning method for deep visual tracking, which selects and annotates the unlabeled samples to train the deep CNNs model.
Under the guidance of active learning, the tracker based on the trained deep CNNs model can achieve competitive tracking performance while reducing the labeling cost.
arXiv Detail & Related papers (2021-10-17T11:47:56Z) - Semi-supervised Long-tailed Recognition using Alternate Sampling [95.93760490301395]
Main challenges in long-tailed recognition come from the imbalanced data distribution and sample scarcity in its tail classes.
We propose a new recognition setting, namely semi-supervised long-tailed recognition.
We demonstrate significant accuracy improvements over other competitive methods on two datasets.
arXiv Detail & Related papers (2021-05-01T00:43:38Z) - One Thing One Click: A Self-Training Approach for Weakly Supervised 3D
Semantic Segmentation [78.36781565047656]
We propose "One Thing One Click," meaning that the annotator only needs to label one point per object.
We iteratively conduct the training and label propagation, facilitated by a graph propagation module.
Our results are also comparable to those of the fully supervised counterparts.
arXiv Detail & Related papers (2021-04-06T02:27:25Z) - Deep Semi-supervised Knowledge Distillation for Overlapping Cervical
Cell Instance Segmentation [54.49894381464853]
We propose to leverage both labeled and unlabeled data for instance segmentation with improved accuracy by knowledge distillation.
We propose a novel Mask-guided Mean Teacher framework with Perturbation-sensitive Sample Mining.
Experiments show that the proposed method improves the performance significantly compared with the supervised method learned from labeled data only.
arXiv Detail & Related papers (2020-07-21T13:27:09Z) - DivideMix: Learning with Noisy Labels as Semi-supervised Learning [111.03364864022261]
We propose DivideMix, a framework for learning with noisy labels.
Experiments on multiple benchmark datasets demonstrate substantial improvements over state-of-the-art methods.
arXiv Detail & Related papers (2020-02-18T06:20:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.