Prototype Fission: Closing Set for Robust Open-set Semi-supervised
Learning
- URL: http://arxiv.org/abs/2308.15575v1
- Date: Tue, 29 Aug 2023 19:04:42 GMT
- Title: Prototype Fission: Closing Set for Robust Open-set Semi-supervised
Learning
- Authors: Xuwei Tan, Yi-Jie Huang, Yaqian Li
- Abstract summary: Semi-supervised Learning (SSL) has been proven vulnerable to out-of-distribution (OOD) samples in realistic large-scale unsupervised datasets.
We propose Prototype Fission(PF) to divide class-wise latent spaces into compact sub-spaces by automatic fine-grained latent space mining.
- Score: 6.645479471664253
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Semi-supervised Learning (SSL) has been proven vulnerable to
out-of-distribution (OOD) samples in realistic large-scale unsupervised
datasets due to over-confident pseudo-labeling OODs as in-distribution (ID). A
key underlying problem is class-wise latent space spreading from closed seen
space to open unseen space, and the bias is further magnified in SSL's
self-training loops. To close the ID distribution set so that OODs are better
rejected for safe SSL, we propose Prototype Fission(PF) to divide class-wise
latent spaces into compact sub-spaces by automatic fine-grained latent space
mining, driven by coarse-grained labels only. Specifically, we form multiple
unique learnable sub-class prototypes for each class, optimized towards both
diversity and consistency. The Diversity Modeling term encourages samples to be
clustered by one of the multiple sub-class prototypes, while the Consistency
Modeling term clusters all samples of the same class to a global prototype.
Instead of "opening set", i.e., modeling OOD distribution, Prototype Fission
"closes set" and makes it hard for OOD samples to fit in sub-class latent
space. Therefore, PF is compatible with existing methods for further
performance gains. Extensive experiments validate the effectiveness of our
method in open-set SSL settings in terms of successfully forming sub-classes,
discriminating OODs from IDs and improving overall accuracy. Codes will be
released.
Related papers
- Generalized Semi-Supervised Learning via Self-Supervised Feature Adaptation [87.17768598044427]
Traditional semi-supervised learning assumes that the feature distributions of labeled and unlabeled data are consistent.
We propose Self-Supervised Feature Adaptation (SSFA), a generic framework for improving SSL performance when labeled and unlabeled data come from different distributions.
Our proposed SSFA is applicable to various pseudo-label-based SSL learners and significantly improves performance in labeled, unlabeled, and even unseen distributions.
arXiv Detail & Related papers (2024-05-31T03:13:45Z) - Exploration and Exploitation of Unlabeled Data for Open-Set
Semi-Supervised Learning [130.56124475528475]
We address a complex but practical scenario in semi-supervised learning (SSL) named open-set SSL, where unlabeled data contain both in-distribution (ID) and out-of-distribution (OOD) samples.
Our proposed method achieves state-of-the-art in several challenging benchmarks, and improves upon existing SSL methods even when ID samples are totally absent in unlabeled data.
arXiv Detail & Related papers (2023-06-30T14:25:35Z) - Latent Distribution Adjusting for Face Anti-Spoofing [29.204168516602568]
We propose a unified framework called Latent Distribution Adjusting (LDA) to improve the robustness of the face anti-spoofing (FAS) model.
To enhance the intra-class compactness and inter-class discrepancy, we propose a margin-based loss for providing distribution constrains for prototype learning.
Our framework can 1) make the final representation space both intra-class compact and inter-class separable, 2) outperform the state-of-the-art methods on multiple standard FAS benchmarks.
arXiv Detail & Related papers (2023-05-16T08:43:14Z) - OpenLDN: Learning to Discover Novel Classes for Open-World
Semi-Supervised Learning [110.40285771431687]
Semi-supervised learning (SSL) is one of the dominant approaches to address the annotation bottleneck of supervised learning.
Recent SSL methods can effectively leverage a large repository of unlabeled data to improve performance while relying on a small set of labeled data.
This work introduces OpenLDN that utilizes a pairwise similarity loss to discover novel classes.
arXiv Detail & Related papers (2022-07-05T18:51:05Z) - Deep Low-Density Separation for Semi-Supervised Classification [0.0]
We introduce a novel hybrid method that applies low-density separation to the embedded features.
Our approach effectively classifies thousands of unlabeled users from a relatively small number of hand-classified examples.
arXiv Detail & Related papers (2022-05-22T11:00:55Z) - Trash to Treasure: Harvesting OOD Data with Cross-Modal Matching for
Open-Set Semi-Supervised Learning [101.28281124670647]
Open-set semi-supervised learning (open-set SSL) investigates a challenging but practical scenario where out-of-distribution (OOD) samples are contained in the unlabeled data.
We propose a novel training mechanism that could effectively exploit the presence of OOD data for enhanced feature learning.
Our approach substantially lifts the performance on open-set SSL and outperforms the state-of-the-art by a large margin.
arXiv Detail & Related papers (2021-08-12T09:14:44Z) - Multi-Class Data Description for Out-of-distribution Detection [25.853322158250435]
Deep-MCDD is effective to detect out-of-distribution (OOD) samples as well as classify in-distribution (ID) samples.
By integrating the concept of Gaussian discriminant analysis into deep neural networks, we propose a deep learning objective to learn class-conditional distributions.
arXiv Detail & Related papers (2021-04-02T08:41:51Z) - Distribution Aligning Refinery of Pseudo-label for Imbalanced
Semi-supervised Learning [126.31716228319902]
We develop Distribution Aligning Refinery of Pseudo-label (DARP) algorithm.
We show that DARP is provably and efficiently compatible with state-of-the-art SSL schemes.
arXiv Detail & Related papers (2020-07-17T09:16:05Z) - From Anchor Generation to Distribution Alignment: Learning a
Discriminative Embedding Space for Zero-Shot Recognition [46.47620562161315]
In zero-shot learning (ZSL), the samples to be classified are usually projected into side information templates such as attributes.
We propose a novel framework called Discriminative Anchor Generation and Distribution Alignment Model (DAGDA)
Firstly, in order to rectify the distribution of original templates, a diffusion based graph convolutional network, which can explicitly model the interaction between class and side information, is proposed to produce discriminative anchors.
Secondly, to further align the samples with the corresponding anchors in anchor space, which aims to refine the distribution in a fine-grained manner, we introduce a semantic relation regularization
arXiv Detail & Related papers (2020-02-10T05:25:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.