KD-FixMatch: Knowledge Distillation Siamese Neural Networks
- URL: http://arxiv.org/abs/2309.05826v1
- Date: Mon, 11 Sep 2023 21:11:48 GMT
- Title: KD-FixMatch: Knowledge Distillation Siamese Neural Networks
- Authors: Chien-Chih Wang, Shaoyuan Xu, Jinmiao Fu, Yang Liu, Bryan Wang
- Abstract summary: KD-FixMatch is a novel SSL algorithm that addresses the limitations of FixMatch by incorporating knowledge distillation.
The algorithm utilizes a combination of sequential and simultaneous training of SNNs to enhance performance and reduce performance degradation.
Our results indicate that KD-FixMatch has a better training starting point that leads to improved model performance compared to FixMatch.
- Score: 13.678635878305247
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Semi-supervised learning (SSL) has become a crucial approach in deep learning
as a way to address the challenge of limited labeled data. The success of deep
neural networks heavily relies on the availability of large-scale high-quality
labeled data. However, the process of data labeling is time-consuming and
unscalable, leading to shortages in labeled data. SSL aims to tackle this
problem by leveraging additional unlabeled data in the training process. One of
the popular SSL algorithms, FixMatch, trains identical weight-sharing teacher
and student networks simultaneously using a siamese neural network (SNN).
However, it is prone to performance degradation when the pseudo labels are
heavily noisy in the early training stage. We present KD-FixMatch, a novel SSL
algorithm that addresses the limitations of FixMatch by incorporating knowledge
distillation. The algorithm utilizes a combination of sequential and
simultaneous training of SNNs to enhance performance and reduce performance
degradation. Firstly, an outer SNN is trained using labeled and unlabeled data.
After that, the network of the well-trained outer SNN generates pseudo labels
for the unlabeled data, from which a subset of unlabeled data with trusted
pseudo labels is then carefully created through high-confidence sampling and
deep embedding clustering. Finally, an inner SNN is trained with the labeled
data, the unlabeled data, and the subset of unlabeled data with trusted pseudo
labels. Experiments on four public data sets demonstrate that KD-FixMatch
outperforms FixMatch in all cases. Our results indicate that KD-FixMatch has a
better training starting point that leads to improved model performance
compared to FixMatch.
Related papers
- ProtoCon: Pseudo-label Refinement via Online Clustering and Prototypical
Consistency for Efficient Semi-supervised Learning [60.57998388590556]
ProtoCon is a novel method for confidence-based pseudo-labeling.
Online nature of ProtoCon allows it to utilise the label history of the entire dataset in one training cycle.
It delivers significant gains and faster convergence over state-of-the-art datasets.
arXiv Detail & Related papers (2023-03-22T23:51:54Z) - NP-Match: Towards a New Probabilistic Model for Semi-Supervised Learning [86.60013228560452]
Semi-supervised learning (SSL) has been widely explored in recent years, and it is an effective way of leveraging unlabeled data.
In this work, we adjust neural processes (NPs) to the semi-supervised image classification task, resulting in a new method named NP-Match.
NP-Match implicitly compares data points when making predictions, and as a result, the prediction of each unlabeled data point is affected by the labeled data points.
arXiv Detail & Related papers (2023-01-31T11:44:45Z) - NP-Match: When Neural Processes meet Semi-Supervised Learning [133.009621275051]
Semi-supervised learning (SSL) has been widely explored in recent years, and it is an effective way of leveraging unlabeled data to reduce the reliance on labeled data.
In this work, we adjust neural processes (NPs) to the semi-supervised image classification task, resulting in a new method named NP-Match.
arXiv Detail & Related papers (2022-07-03T15:24:31Z) - CLS: Cross Labeling Supervision for Semi-Supervised Learning [9.929229055862491]
Cross Labeling Supervision ( CLS) is a framework that generalizes the typical pseudo-labeling process.
CLS allows the creation of both pseudo and complementary labels to support both positive and negative learning.
arXiv Detail & Related papers (2022-02-17T08:09:40Z) - Robust Deep Semi-Supervised Learning: A Brief Introduction [63.09703308309176]
Semi-supervised learning (SSL) aims to improve learning performance by leveraging unlabeled data when labels are insufficient.
SSL with deep models has proven to be successful on standard benchmark tasks.
However, they are still vulnerable to various robustness threats in real-world applications.
arXiv Detail & Related papers (2022-02-12T04:16:41Z) - Boosting Semi-Supervised Face Recognition with Noise Robustness [54.342992887966616]
This paper presents an effective solution to semi-supervised face recognition that is robust to the label noise aroused by the auto-labelling.
We develop a semi-supervised face recognition solution, named Noise Robust Learning-Labelling (NRoLL), which is based on the robust training ability empowered by GN.
arXiv Detail & Related papers (2021-05-10T14:43:11Z) - Relieving the Plateau: Active Semi-Supervised Learning for a Better
Landscape [2.3046646540823916]
Semi-supervised learning (SSL) leverages unlabeled data that are more accessible than their labeled counterparts.
Active learning (AL) selects unlabeled instances to be annotated by a human-in-the-loop in hopes of better performance with less labeled data.
We propose convergence rate control (CRC), an AL algorithm that selects unlabeled data to improve the problem conditioning upon inclusion to the labeled set.
arXiv Detail & Related papers (2021-04-08T06:03:59Z) - KNN-enhanced Deep Learning Against Noisy Labels [4.765948508271371]
Supervised learning on Deep Neural Networks (DNNs) is data hungry.
In this work, we propose to apply deep KNN for label cleanup.
We iteratively train the neural network and update labels to simultaneously proceed towards higher label recovery rate and better classification performance.
arXiv Detail & Related papers (2020-12-08T05:21:29Z) - Self-Competitive Neural Networks [0.0]
Deep Neural Networks (DNNs) have improved the accuracy of classification problems in lots of applications.
One of the challenges in training a DNN is its need to be fed by an enriched dataset to increase its accuracy and avoid it suffering from overfitting.
Recently, researchers have worked extensively to propose methods for data augmentation.
In this paper, we generate adversarial samples to refine the Domains of Attraction (DoAs) of each class. In this approach, at each stage, we use the model learned by the primary and generated adversarial data (up to that stage) to manipulate the primary data in a way that look complicated to
arXiv Detail & Related papers (2020-08-22T12:28:35Z) - Temporal Calibrated Regularization for Robust Noisy Label Learning [60.90967240168525]
Deep neural networks (DNNs) exhibit great success on many tasks with the help of large-scale well annotated datasets.
However, labeling large-scale data can be very costly and error-prone so that it is difficult to guarantee the annotation quality.
We propose a Temporal Calibrated Regularization (TCR) in which we utilize the original labels and the predictions in the previous epoch together.
arXiv Detail & Related papers (2020-07-01T04:48:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.