High-Dimensional Bayesian Optimization via Semi-Supervised Learning with
Optimized Unlabeled Data Sampling
- URL: http://arxiv.org/abs/2305.02614v3
- Date: Sat, 3 Feb 2024 16:23:06 GMT
- Title: High-Dimensional Bayesian Optimization via Semi-Supervised Learning with
Optimized Unlabeled Data Sampling
- Authors: Yuxuan Yin, Yu Wang and Peng Li
- Abstract summary: $texttTSBO$ incorporates a teacher model, an unlabeled data sampler, and a student model.
The student is trained on unlabeled data locations generated by the sampler, with pseudo labels predicted by the teacher.
$texttTSBO$ demonstrates significantly improved sample-efficiency in several global optimization tasks under tight labeled data budgets.
- Score: 6.927830939687371
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce a novel semi-supervised learning approach, named Teacher-Student
Bayesian Optimization ($\texttt{TSBO}$), integrating the teacher-student
paradigm into BO to minimize expensive labeled data queries for the first time.
$\texttt{TSBO}$ incorporates a teacher model, an unlabeled data sampler, and a
student model. The student is trained on unlabeled data locations generated by
the sampler, with pseudo labels predicted by the teacher. The interplay between
these three components implements a unique selective regularization to the
teacher in the form of student feedback. This scheme enables the teacher to
predict high-quality pseudo labels, enhancing the generalization of the GP
surrogate model in the search space. To fully exploit $\texttt{TSBO}$, we
propose two optimized unlabeled data samplers to construct effective student
feedback that well aligns with the objective of Bayesian optimization.
Furthermore, we quantify and leverage the uncertainty of the teacher-student
model for the provision of reliable feedback to the teacher in the presence of
risky pseudo-label predictions. $\texttt{TSBO}$ demonstrates significantly
improved sample-efficiency in several global optimization tasks under tight
labeled data budgets.
Related papers
- CoDTS: Enhancing Sparsely Supervised Collaborative Perception with a Dual Teacher-Student Framework [15.538850922083652]
We propose an end-to-end Collaborative perception Dual Teacher-Student framework (CoDTS)
It employs adaptive complementary learning to produce both high-quality and high-quantity pseudo labels.
CoDTS effectively ensures an optimal balance of pseudo labels in both quality and quantity.
arXiv Detail & Related papers (2024-12-11T12:34:37Z) - Semi-Supervised Semantic Segmentation via Gentle Teaching Assistant [72.4512562104361]
We argue that the unlabeled data with pseudo labels can facilitate the learning of representative features in the feature extractor.
Motivated by this consideration, we propose a novel framework, Gentle Teaching Assistant (GTA-Seg) to disentangle the effects of pseudo labels on feature extractor and mask predictor.
arXiv Detail & Related papers (2023-01-18T07:11:24Z) - Enhancing Self-Training Methods [0.0]
Semi-supervised learning approaches train on small sets of labeled data along with large sets of unlabeled data.
Self-training is a semi-supervised teacher-student approach that often suffers from the problem of "confirmation bias"
arXiv Detail & Related papers (2023-01-18T03:56:17Z) - Non-iterative optimization of pseudo-labeling thresholds for training
object detection models from multiple datasets [2.1485350418225244]
We propose a non-iterative method to optimize pseudo-labeling thresholds for learning object detection from a collection of low-cost datasets.
We experimentally demonstrate that our proposed method achieves an mAP comparable to that of grid search on the COCO and VOC datasets.
arXiv Detail & Related papers (2022-10-19T00:31:34Z) - Label Matching Semi-Supervised Object Detection [85.99282969977541]
Semi-supervised object detection has made significant progress with the development of mean teacher driven self-training.
Label mismatch problem is not yet fully explored in the previous works, leading to severe confirmation bias during self-training.
We propose a simple yet effective LabelMatch framework from two different yet complementary perspectives.
arXiv Detail & Related papers (2022-06-14T05:59:41Z) - SLADE: A Self-Training Framework For Distance Metric Learning [75.54078592084217]
We present a self-training framework, SLADE, to improve retrieval performance by leveraging additional unlabeled data.
We first train a teacher model on the labeled data and use it to generate pseudo labels for the unlabeled data.
We then train a student model on both labels and pseudo labels to generate final feature embeddings.
arXiv Detail & Related papers (2020-11-20T08:26:10Z) - Semi-Automatic Data Annotation guided by Feature Space Projection [117.9296191012968]
We present a semi-automatic data annotation approach based on suitable feature space projection and semi-supervised label estimation.
We validate our method on the popular MNIST dataset and on images of human intestinal parasites with and without fecal impurities.
Our results demonstrate the added-value of visual analytics tools that combine complementary abilities of humans and machines for more effective machine learning.
arXiv Detail & Related papers (2020-07-27T17:03:50Z) - Deep Semi-supervised Knowledge Distillation for Overlapping Cervical
Cell Instance Segmentation [54.49894381464853]
We propose to leverage both labeled and unlabeled data for instance segmentation with improved accuracy by knowledge distillation.
We propose a novel Mask-guided Mean Teacher framework with Perturbation-sensitive Sample Mining.
Experiments show that the proposed method improves the performance significantly compared with the supervised method learned from labeled data only.
arXiv Detail & Related papers (2020-07-21T13:27:09Z) - ATSO: Asynchronous Teacher-Student Optimization for Semi-Supervised
Medical Image Segmentation [99.90263375737362]
We propose ATSO, an asynchronous version of teacher-student optimization.
ATSO partitions the unlabeled data into two subsets and alternately uses one subset to fine-tune the model and updates the label on the other subset.
We evaluate ATSO on two popular medical image segmentation datasets and show its superior performance in various semi-supervised settings.
arXiv Detail & Related papers (2020-06-24T04:05:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.