Enhancing Sample Utilization through Sample Adaptive Augmentation in
Semi-Supervised Learning
- URL: http://arxiv.org/abs/2309.03598v1
- Date: Thu, 7 Sep 2023 09:50:45 GMT
- Title: Enhancing Sample Utilization through Sample Adaptive Augmentation in
Semi-Supervised Learning
- Authors: Guan Gui, Zhen Zhao, Lei Qi, Luping Zhou, Lei Wang, Yinghuan Shi
- Abstract summary: In semi-supervised learning, unlabeled samples can be utilized through augmentation and consistency regularization.
Existing SSL models overlook the characteristics of naive samples, and they just apply the same learning strategy to all samples.
We propose Sample adaptive augmentation (SAA) to give attention to naive samples and augmenting them in a more diverse manner.
- Score: 47.677929366323596
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In semi-supervised learning, unlabeled samples can be utilized through
augmentation and consistency regularization. However, we observed certain
samples, even undergoing strong augmentation, are still correctly classified
with high confidence, resulting in a loss close to zero. It indicates that
these samples have been already learned well and do not provide any additional
optimization benefits to the model. We refer to these samples as ``naive
samples". Unfortunately, existing SSL models overlook the characteristics of
naive samples, and they just apply the same learning strategy to all samples.
To further optimize the SSL model, we emphasize the importance of giving
attention to naive samples and augmenting them in a more diverse manner. Sample
adaptive augmentation (SAA) is proposed for this stated purpose and consists of
two modules: 1) sample selection module; 2) sample augmentation module.
Specifically, the sample selection module picks out {naive samples} based on
historical training information at each epoch, then the naive samples will be
augmented in a more diverse manner in the sample augmentation module. Thanks to
the extreme ease of implementation of the above modules, SAA is advantageous
for being simple and lightweight. We add SAA on top of FixMatch and FlexMatch
respectively, and experiments demonstrate SAA can significantly improve the
models. For example, SAA helped improve the accuracy of FixMatch from 92.50% to
94.76% and that of FlexMatch from 95.01% to 95.31% on CIFAR-10 with 40 labels.
Related papers
- Enhancing Sample Selection by Cutting Mislabeled Easy Examples [62.13094877228772]
We show that mislabeled examples correctly predicted by the model early in the training process are particularly harmful to model performance.
We propose Early Cutting, which employs the model's later training state to re-select the confident subset identified early in training.
arXiv Detail & Related papers (2025-02-12T09:12:45Z) - Adaptive Inference-Time Compute: LLMs Can Predict if They Can Do Better, Even Mid-Generation [51.127054971591924]
We introduce a new generative self-evaluation scheme designed to adaptively reduce the number of generated samples.
We demonstrate that 74% of the improvement from using 16 samples can be achieved with only 1.2 samples on average.
arXiv Detail & Related papers (2024-10-03T17:47:29Z) - Large Language Monkeys: Scaling Inference Compute with Repeated Sampling [81.34900892130929]
We explore inference compute as another axis for scaling, using the simple technique of repeatedly sampling candidate solutions from a model.
Across multiple tasks and models, we observe that coverage scales with the number of samples over four orders of magnitude.
In domains like coding and formal proofs, where answers can be automatically verified, these increases in coverage directly translate into improved performance.
arXiv Detail & Related papers (2024-07-31T17:57:25Z) - Uncertainty Aware Learning for Language Model Alignment [97.36361196793929]
We propose uncertainty-aware learning (UAL) to improve the model alignment of different task scenarios.
We implement UAL in a simple fashion -- adaptively setting the label smoothing value of training according to the uncertainty of individual samples.
Experiments on widely used benchmarks demonstrate that our UAL significantly and consistently outperforms standard supervised fine-tuning.
arXiv Detail & Related papers (2024-06-07T11:37:45Z) - Rethinking Samples Selection for Contrastive Learning: Mining of
Potential Samples [5.586563813796839]
Contrastive learning predicts whether two images belong to the same category by training a model to make their feature representations as close or as far away as possible.
We take into account both positive and negative samples, and mining potential samples from two aspects.
Our method achieves 88.57%, 61.10%, and 36.69% top-1 accuracy on CIFAR10, CIFAR100, and TinyImagenet, respectively.
arXiv Detail & Related papers (2023-11-01T08:08:06Z) - Sampling Through the Lens of Sequential Decision Making [9.101505546901999]
We propose a reward-guided sampling strategy called Adaptive Sample with Reward (ASR)
Our approach optimally adjusts the sampling process to achieve optimal performance.
Empirical results in information retrieval and clustering demonstrate ASR's superb performance across different datasets.
arXiv Detail & Related papers (2022-08-17T04:01:29Z) - ReSmooth: Detecting and Utilizing OOD Samples when Training with Data
Augmentation [57.38418881020046]
Recent DA techniques always meet the need for diversity in augmented training samples.
An augmentation strategy that has a high diversity usually introduces out-of-distribution (OOD) augmented samples.
We propose ReSmooth, a framework that firstly detects OOD samples in augmented samples and then leverages them.
arXiv Detail & Related papers (2022-05-25T09:29:27Z) - Learning Fast Samplers for Diffusion Models by Differentiating Through
Sample Quality [44.37533757879762]
We introduce Differentiable Diffusion Sampler Search (DDSS), a method that optimize fast samplers for any pre-trained diffusion model.
We also present Generalized Gaussian Diffusion Models (GGDM), a family of flexible non-Markovian samplers for diffusion models.
Our method is compatible with any pre-trained diffusion model without fine-tuning or re-training required.
arXiv Detail & Related papers (2022-02-11T18:53:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.