A Probabilistic Model for Discriminative and Neuro-Symbolic
Semi-Supervised Learning
- URL: http://arxiv.org/abs/2006.05896v4
- Date: Mon, 31 May 2021 12:20:00 GMT
- Title: A Probabilistic Model for Discriminative and Neuro-Symbolic
Semi-Supervised Learning
- Authors: Carl Allen, Ivana Bala\v{z}evi\'c, Timothy Hospedales
- Abstract summary: We present a probabilistic model for discriminative SSL, that mirrors its classical generative counterpart.
We show several well-known SSL methods can be interpreted as approximating this prior, and can be improved upon.
We extend the discriminative model to neuro-symbolic SSL, where label features satisfy logical rules, by showing such rules relate directly to the above prior.
- Score: 6.789370732159177
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Much progress has been made in semi-supervised learning (SSL) by combining
methods that exploit different aspects of the data distribution, e.g.
consistency regularisation relies on properties of $p(x)$, whereas entropy
minimisation pertains to the label distribution $p(y|x)$. Focusing on the
latter, we present a probabilistic model for discriminative SSL, that mirrors
its classical generative counterpart. Under the assumption $y|x$ is
deterministic, the prior over latent variables becomes discrete. We show that
several well-known SSL methods can be interpreted as approximating this prior,
and can be improved upon. We extend the discriminative model to neuro-symbolic
SSL, where label features satisfy logical rules, by showing such rules relate
directly to the above prior, thus justifying a family of methods that link
statistical learning and logical reasoning, and unifying them with regular SSL.
Related papers
- A Probabilistic Model for Self-Supervised Learning [6.178817969919849]
Self-supervised learning (SSL) aims to find meaningful representations from unlabeled data by encoding semantic similarities through data augmentations.
It is not yet known whether commonly used SSL loss functions can be related to a statistical model.
We consider a latent variable statistical model for SSL that exhibits an interesting property: Depending on the informativeness of the data augmentations, the MLE of the model either reduces to PCA, or approaches a simple non-contrastive loss.
arXiv Detail & Related papers (2025-01-22T17:25:47Z) - Explicit Mutual Information Maximization for Self-Supervised Learning [23.41734709882332]
Theoretically, mutual information (MIM) is an optimal criterion for self-supervised learning (SSL)
This work shows that, based on the explicit in property of MI, MI can be applied to SSL under a generic distribution assumption.
We derive a loss function based on the MIM criterion using only second-order statistics.
arXiv Detail & Related papers (2024-09-07T07:44:33Z) - Semi-Supervised Sparse Gaussian Classification: Provable Benefits of Unlabeled Data [6.812609988733991]
We study SSL for high dimensional Gaussian classification.
We analyze information theoretic lower bounds for accurate feature selection.
We present simulations that complement our theoretical analysis.
arXiv Detail & Related papers (2024-09-05T08:21:05Z) - A Closer Look at Benchmarking Self-Supervised Pre-training with Image Classification [51.35500308126506]
Self-supervised learning (SSL) is a machine learning approach where the data itself provides supervision, eliminating the need for external labels.
We study how classification-based evaluation protocols for SSL correlate and how well they predict downstream performance on different dataset types.
arXiv Detail & Related papers (2024-07-16T23:17:36Z) - Rebalanced Zero-shot Learning [18.52913434532522]
Zero-shot learning (ZSL) aims to identify unseen classes with zero samples during training.
We introduce an imbalanced learning framework into ZSL.
We then propose a re-weighted loss termed Re-balanced Mean-Squared Error (ReMSE)
arXiv Detail & Related papers (2022-10-13T13:43:40Z) - OpenLDN: Learning to Discover Novel Classes for Open-World
Semi-Supervised Learning [110.40285771431687]
Semi-supervised learning (SSL) is one of the dominant approaches to address the annotation bottleneck of supervised learning.
Recent SSL methods can effectively leverage a large repository of unlabeled data to improve performance while relying on a small set of labeled data.
This work introduces OpenLDN that utilizes a pairwise similarity loss to discover novel classes.
arXiv Detail & Related papers (2022-07-05T18:51:05Z) - A Strong Baseline for Semi-Supervised Incremental Few-Shot Learning [54.617688468341704]
Few-shot learning aims to learn models that generalize to novel classes with limited training samples.
We propose a novel paradigm containing two parts: (1) a well-designed meta-training algorithm for mitigating ambiguity between base and novel classes caused by unreliable pseudo labels and (2) a model adaptation mechanism to learn discriminative features for novel classes while preserving base knowledge using few labeled and all the unlabeled data.
arXiv Detail & Related papers (2021-10-21T13:25:52Z) - Self-supervised Learning is More Robust to Dataset Imbalance [65.84339596595383]
We investigate self-supervised learning under dataset imbalance.
Off-the-shelf self-supervised representations are already more robust to class imbalance than supervised representations.
We devise a re-weighted regularization technique that consistently improves the SSL representation quality on imbalanced datasets.
arXiv Detail & Related papers (2021-10-11T06:29:56Z) - On Data-Augmentation and Consistency-Based Semi-Supervised Learning [77.57285768500225]
Recently proposed consistency-based Semi-Supervised Learning (SSL) methods have advanced the state of the art in several SSL tasks.
Despite these advances, the understanding of these methods is still relatively limited.
arXiv Detail & Related papers (2021-01-18T10:12:31Z) - Semi-supervised learning objectives as log-likelihoods in a generative
model of data curation [32.45282187405337]
We formulate SSL objectives as a log-likelihood in a generative model of data curation.
We give a proof-of-principle for Bayesian SSL on toy data.
arXiv Detail & Related papers (2020-08-13T13:50:27Z) - Distribution Aligning Refinery of Pseudo-label for Imbalanced
Semi-supervised Learning [126.31716228319902]
We develop Distribution Aligning Refinery of Pseudo-label (DARP) algorithm.
We show that DARP is provably and efficiently compatible with state-of-the-art SSL schemes.
arXiv Detail & Related papers (2020-07-17T09:16:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.