Revisiting Deep Semi-supervised Learning: An Empirical Distribution
Alignment Framework and Its Generalization Bound
- URL: http://arxiv.org/abs/2203.06639v1
- Date: Sun, 13 Mar 2022 11:59:52 GMT
- Title: Revisiting Deep Semi-supervised Learning: An Empirical Distribution
Alignment Framework and Its Generalization Bound
- Authors: Feiyu Wang, Qin Wang, Wen Li, Dong Xu, Luc Van Gool
- Abstract summary: We propose a new deep semi-supervised learning framework called Semi-supervised Learning by Empirical Distribution Alignment (SLEDA)
We show the generalization error of semi-supervised learning can be effectively bounded by minimizing the training error on labeled data.
Building upon our new framework and the theoretical bound, we develop a simple and effective deep semi-supervised learning method called Augmented Distribution Alignment Network (ADA-Net)
- Score: 97.93945601881407
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work, we revisit the semi-supervised learning (SSL) problem from a
new perspective of explicitly reducing empirical distribution mismatch between
labeled and unlabeled samples. Benefited from this new perspective, we first
propose a new deep semi-supervised learning framework called Semi-supervised
Learning by Empirical Distribution Alignment (SLEDA), in which existing
technologies from the domain adaptation community can be readily used to
address the semi-supervised learning problem through reducing the empirical
distribution distance between labeled and unlabeled data. Based on this
framework, we also develop a new theoretical generalization bound for the
research community to better understand the semi-supervised learning problem,
in which we show the generalization error of semi-supervised learning can be
effectively bounded by minimizing the training error on labeled data and the
empirical distribution distance between labeled and unlabeled data. Building
upon our new framework and the theoretical bound, we develop a simple and
effective deep semi-supervised learning method called Augmented Distribution
Alignment Network (ADA-Net) by simultaneously adopting the well-established
adversarial training strategy from the domain adaptation community and a simple
sample interpolation strategy for data augmentation. Additionally, we
incorporate both strategies in our ADA-Net into two exiting SSL methods to
further improve their generalization capability, which indicates that our new
framework provides a complementary solution for solving the SSL problem. Our
comprehensive experimental results on two benchmark datasets SVHN and CIFAR-10
for the semi-supervised image recognition task and another two benchmark
datasets ModelNet40 and ShapeNet55 for the semi-supervised point cloud
recognition task demonstrate the effectiveness of our proposed framework for
SSL.
Related papers
- Multi-Agent Reinforcement Learning from Human Feedback: Data Coverage and Algorithmic Techniques [65.55451717632317]
We study Multi-Agent Reinforcement Learning from Human Feedback (MARLHF), exploring both theoretical foundations and empirical validations.
We define the task as identifying Nash equilibrium from a preference-only offline dataset in general-sum games.
Our findings underscore the multifaceted approach required for MARLHF, paving the way for effective preference-based multi-agent systems.
arXiv Detail & Related papers (2024-09-01T13:14:41Z) - ItTakesTwo: Leveraging Peer Representations for Semi-supervised LiDAR Semantic Segmentation [24.743048965822297]
This paper introduces a novel semi-supervised LiDAR semantic segmentation framework called ItTakesTwo (IT2)
IT2 is designed to ensure consistent predictions from peer LiDAR representations, thereby improving the perturbation effectiveness in consistency learning.
Results on public benchmarks show that our approach achieves remarkable improvements over the previous state-of-the-art (SOTA) methods in the field.
arXiv Detail & Related papers (2024-07-09T18:26:53Z) - CAusal and collaborative proxy-tasKs lEarning for Semi-Supervised Domain
Adaptation [20.589323508870592]
Semi-supervised domain adaptation (SSDA) adapts a learner to a new domain by effectively utilizing source domain data and a few labeled target samples.
We show that the proposed model significantly outperforms SOTA methods in terms of effectiveness and generalisability on SSDA datasets.
arXiv Detail & Related papers (2023-03-30T16:48:28Z) - Interpolation-based Correlation Reduction Network for Semi-Supervised
Graph Learning [49.94816548023729]
We propose a novel graph contrastive learning method, termed Interpolation-based Correlation Reduction Network (ICRN)
In our method, we improve the discriminative capability of the latent feature by enlarging the margin of decision boundaries.
By combining the two settings, we extract rich supervision information from both the abundant unlabeled nodes and the rare yet valuable labeled nodes for discnative representation learning.
arXiv Detail & Related papers (2022-06-06T14:26:34Z) - Multi-level Consistency Learning for Semi-supervised Domain Adaptation [85.90600060675632]
Semi-supervised domain adaptation (SSDA) aims to apply knowledge learned from a fully labeled source domain to a scarcely labeled target domain.
We propose a Multi-level Consistency Learning framework for SSDA.
arXiv Detail & Related papers (2022-05-09T06:41:18Z) - An Information-theoretical Approach to Semi-supervised Learning under
Covariate-shift [24.061858945664856]
A common assumption in semi-supervised learning is that the labeled, unlabeled, and test data are drawn from the same distribution.
We propose an approach for semi-supervised learning algorithms that is capable of addressing this issue.
Our framework also recovers some popular methods, including entropy minimization and pseudo-labeling.
arXiv Detail & Related papers (2022-02-24T14:25:14Z) - Semi-supervised Domain Adaptive Structure Learning [72.01544419893628]
Semi-supervised domain adaptation (SSDA) is a challenging problem requiring methods to overcome both 1) overfitting towards poorly annotated data and 2) distribution shift across domains.
We introduce an adaptive structure learning method to regularize the cooperation of SSL and DA.
arXiv Detail & Related papers (2021-12-12T06:11:16Z) - Learning Invariant Representations and Risks for Semi-supervised Domain
Adaptation [109.73983088432364]
We propose the first method that aims to simultaneously learn invariant representations and risks under the setting of semi-supervised domain adaptation (Semi-DA)
We introduce the LIRR algorithm for jointly textbfLearning textbfInvariant textbfRepresentations and textbfRisks.
arXiv Detail & Related papers (2020-10-09T15:42:35Z) - Consistency Regularization with Generative Adversarial Networks for
Semi-Supervised Learning [2.9707483702447783]
Adversarial Adversarial Networks (GANs) based semi-supervised learning (SSL) approaches are shown to improve classification performance by utilizing a large number of unlabeled samples.
However, their performance still lags behind the state-of-the-art non-GAN based SSL approaches.
We identify that the main reason for this is the lack of consistency in class probability predictions on the same image under local perturbations.
arXiv Detail & Related papers (2020-07-08T01:47:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.