Perturbed and Strict Mean Teachers for Semi-supervised Semantic
Segmentation
- URL: http://arxiv.org/abs/2111.12903v1
- Date: Thu, 25 Nov 2021 04:30:56 GMT
- Title: Perturbed and Strict Mean Teachers for Semi-supervised Semantic
Segmentation
- Authors: Yuyuan Liu and Yu Tian and Yuanhong Chen and Fengbei Liu and Vasileios
Belagiannis and Gustavo Carneiro
- Abstract summary: In this paper, we address the prediction accuracy problem of consistency learning methods with novel extensions of the mean-teacher (MT) model.
The accurate prediction by this model allows us to use a challenging combination of network, input data and feature perturbations to improve the consistency learning generalisation.
Results on public benchmarks show that our approach achieves remarkable improvements over the previous SOTA methods in the field.
- Score: 22.5935068122522
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Consistency learning using input image, feature, or network perturbations has
shown remarkable results in semi-supervised semantic segmentation, but this
approach can be seriously affected by inaccurate predictions of unlabelled
training images. There are two consequences of these inaccurate predictions: 1)
the training based on the ``strict'' cross-entropy (CE) loss can easily overfit
prediction mistakes, leading to confirmation bias; and 2) the perturbations
applied to these inaccurate predictions will use potentially erroneous
predictions as training signals, degrading consistency learning. In this paper,
we address the prediction accuracy problem of consistency learning methods with
novel extensions of the mean-teacher (MT) model, which include a new auxiliary
teacher, and the replacement of MT's mean square error (MSE) by a stricter
confidence-weighted cross-entropy (Conf-CE) loss. The accurate prediction by
this model allows us to use a challenging combination of network, input data
and feature perturbations to improve the consistency learning generalisation,
where the feature perturbations consist of a new adversarial perturbation.
Results on public benchmarks show that our approach achieves remarkable
improvements over the previous SOTA methods in the field.
Related papers
- Towards Robust and Interpretable EMG-based Hand Gesture Recognition using Deep Metric Meta Learning [37.21211404608413]
We propose a shift to deep metric-based meta-learning in EMG PR to supervise the creation of meaningful and interpretable representations.
We derive a robust class proximity-based confidence estimator that leads to a better rejection of incorrect decisions.
arXiv Detail & Related papers (2024-04-17T23:37:50Z) - Leveraging Unlabeled Data for 3D Medical Image Segmentation through
Self-Supervised Contrastive Learning [3.7395287262521717]
Current 3D semi-supervised segmentation methods face significant challenges such as limited consideration of contextual information.
We introduce two distinctworks designed to explore and exploit the discrepancies between them, ultimately correcting the erroneous prediction results.
We employ a self-supervised contrastive learning paradigm to distinguish between reliable and unreliable predictions.
arXiv Detail & Related papers (2023-11-21T14:03:16Z) - Learning Sample Difficulty from Pre-trained Models for Reliable
Prediction [55.77136037458667]
We propose to utilize large-scale pre-trained models to guide downstream model training with sample difficulty-aware entropy regularization.
We simultaneously improve accuracy and uncertainty calibration across challenging benchmarks.
arXiv Detail & Related papers (2023-04-20T07:29:23Z) - Self-Correctable and Adaptable Inference for Generalizable Human Pose
Estimation [25.14459820592431]
We introduce a self-correctable and adaptable inference (SCAI) method to address the generalization challenge of network prediction.
We show that the proposed SCAI method is able to significantly improve the generalization capability and performance of human pose estimation.
arXiv Detail & Related papers (2023-03-20T15:05:24Z) - Feature Perturbation Augmentation for Reliable Evaluation of Importance
Estimators in Neural Networks [5.439020425819001]
Post-hoc interpretability methods attempt to make the inner workings of deep neural networks more interpretable.
One of the most popular evaluation frameworks is to perturb features deemed important by an interpretability method.
We propose feature perturbation augmentation (FPA) which creates and adds perturbed images during the model training.
arXiv Detail & Related papers (2023-03-02T19:05:46Z) - Improving Adaptive Conformal Prediction Using Self-Supervised Learning [72.2614468437919]
We train an auxiliary model with a self-supervised pretext task on top of an existing predictive model and use the self-supervised error as an additional feature to estimate nonconformity scores.
We empirically demonstrate the benefit of the additional information using both synthetic and real data on the efficiency (width), deficit, and excess of conformal prediction intervals.
arXiv Detail & Related papers (2023-02-23T18:57:14Z) - Trust but Verify: Assigning Prediction Credibility by Counterfactual
Constrained Learning [123.3472310767721]
Prediction credibility measures are fundamental in statistics and machine learning.
These measures should account for the wide variety of models used in practice.
The framework developed in this work expresses the credibility as a risk-fit trade-off.
arXiv Detail & Related papers (2020-11-24T19:52:38Z) - Robust Pre-Training by Adversarial Contrastive Learning [120.33706897927391]
Recent work has shown that, when integrated with adversarial training, self-supervised pre-training can lead to state-of-the-art robustness.
We improve robustness-aware self-supervised pre-training by learning representations consistent under both data augmentations and adversarial perturbations.
arXiv Detail & Related papers (2020-10-26T04:44:43Z) - Accurate and Robust Feature Importance Estimation under Distribution
Shifts [49.58991359544005]
PRoFILE is a novel feature importance estimation method.
We show significant improvements over state-of-the-art approaches, both in terms of fidelity and robustness.
arXiv Detail & Related papers (2020-09-30T05:29:01Z) - Grasping Detection Network with Uncertainty Estimation for
Confidence-Driven Semi-Supervised Domain Adaptation [17.16216430459064]
This paper presents an approach enabling the easy domain adaptation through a novel grasping detection network with confidence-driven semi-supervised learning.
The proposed grasping detection network specially provides a prediction uncertainty estimation mechanism by leveraging on Feature Pyramid Network (FPN), and the mean-teacher semi-supervised learning utilizes such uncertainty information to emphasizing the consistency loss only for those unlabelled data with high confidence.
Our results show that the proposed network can achieve high success rate on the Cornell grasping dataset, and for domain adaptation with very limited data, the confidence-driven mean teacher outperforms the original mean teacher and direct training by more than 10% in evaluation
arXiv Detail & Related papers (2020-08-20T07:42:45Z) - Learning to Predict Error for MRI Reconstruction [67.76632988696943]
We demonstrate that predictive uncertainty estimated by the current methods does not highly correlate with prediction error.
We propose a novel method that estimates the target labels and magnitude of the prediction error in two steps.
arXiv Detail & Related papers (2020-02-13T15:55:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.