GEDI: GEnerative and DIscriminative Training for Self-Supervised
Learning
- URL: http://arxiv.org/abs/2212.13425v1
- Date: Tue, 27 Dec 2022 09:33:50 GMT
- Title: GEDI: GEnerative and DIscriminative Training for Self-Supervised
Learning
- Authors: Emanuele Sansone and Robin Manhaeve
- Abstract summary: We study state-of-the-art self-supervised learning objectives and propose a unified formulation based on likelihood learning.
We refer to this combined framework as GEDI, which stands for GEnerative and DIscriminative training.
We show that GEDI outperforms existing self-supervised learning strategies in terms of clustering performance by a wide margin.
- Score: 3.6804038214708563
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Self-supervised learning is a popular and powerful method for utilizing large
amounts of unlabeled data, for which a wide variety of training objectives have
been proposed in the literature. In this study, we perform a Bayesian analysis
of state-of-the-art self-supervised learning objectives and propose a unified
formulation based on likelihood learning. Our analysis suggests a simple method
for integrating self-supervised learning with generative models, allowing for
the joint training of these two seemingly distinct approaches. We refer to this
combined framework as GEDI, which stands for GEnerative and DIscriminative
training. Additionally, we demonstrate an instantiation of the GEDI framework
by integrating an energy-based model with a cluster-based self-supervised
learning model. Through experiments on synthetic and real-world data, including
SVHN, CIFAR10, and CIFAR100, we show that GEDI outperforms existing
self-supervised learning strategies in terms of clustering performance by a
wide margin. We also demonstrate that GEDI can be integrated into a
neural-symbolic framework to address tasks in the small data regime, where it
can use logical constraints to further improve clustering and classification
performance.
Related papers
- A Unified and General Framework for Continual Learning [58.72671755989431]
Continual Learning (CL) focuses on learning from dynamic and changing data distributions while retaining previously acquired knowledge.
Various methods have been developed to address the challenge of catastrophic forgetting, including regularization-based, Bayesian-based, and memory-replay-based techniques.
This research aims to bridge this gap by introducing a comprehensive and overarching framework that encompasses and reconciles these existing methodologies.
arXiv Detail & Related papers (2024-03-20T02:21:44Z) - A Probabilistic Model Behind Self-Supervised Learning [53.64989127914936]
In self-supervised learning (SSL), representations are learned via an auxiliary task without annotated labels.
We present a generative latent variable model for self-supervised learning.
We show that several families of discriminative SSL, including contrastive methods, induce a comparable distribution over representations.
arXiv Detail & Related papers (2024-02-02T13:31:17Z) - A Bayesian Unification of Self-Supervised Clustering and Energy-Based
Models [11.007541337967027]
We perform a Bayesian analysis of state-of-the-art self-supervised learning objectives.
We show that our objective function allows to outperform existing self-supervised learning strategies.
We also demonstrate that GEDI can be integrated into a neuro-symbolic framework.
arXiv Detail & Related papers (2023-12-30T04:46:16Z) - Holonic Learning: A Flexible Agent-based Distributed Machine Learning
Framework [0.0]
Holonic Learning (HoL) is a collaborative and privacy-focused learning framework designed for training deep learning models.
By leveraging holonic concepts, HoL framework establishes a structured self-similar hierarchy in the learning process.
This paper implements HoloAvg, a special variant of HoL that employs weighted averaging for model aggregation across all holons.
arXiv Detail & Related papers (2023-12-29T12:03:42Z) - Semi-supervised learning made simple with self-supervised clustering [65.98152950607707]
Self-supervised learning models have been shown to learn rich visual representations without requiring human annotations.
We propose a conceptually simple yet empirically powerful approach to turn clustering-based self-supervised methods into semi-supervised learners.
arXiv Detail & Related papers (2023-06-13T01:09:18Z) - Learning Symbolic Representations Through Joint GEnerative and
DIscriminative Training [3.6804038214708563]
GEDI is a Bayesian framework that combines self-supervised learning objectives with likelihood-based generative models.
We demonstrate GEDI outperforms existing self-supervised learning strategies in terms of clustering performance by a significant margin.
arXiv Detail & Related papers (2023-04-22T09:35:51Z) - Federated Learning Aggregation: New Robust Algorithms with Guarantees [63.96013144017572]
Federated learning has been recently proposed for distributed model training at the edge.
This paper presents a complete general mathematical convergence analysis to evaluate aggregation strategies in a federated learning framework.
We derive novel aggregation algorithms which are able to modify their model architecture by differentiating client contributions according to the value of their losses.
arXiv Detail & Related papers (2022-05-22T16:37:53Z) - Edge-assisted Democratized Learning Towards Federated Analytics [67.44078999945722]
We show the hierarchical learning structure of the proposed edge-assisted democratized learning mechanism, namely Edge-DemLearn.
We also validate Edge-DemLearn as a flexible model training mechanism to build a distributed control and aggregation methodology in regions.
arXiv Detail & Related papers (2020-12-01T11:46:03Z) - Self-organizing Democratized Learning: Towards Large-scale Distributed
Learning Systems [71.14339738190202]
democratized learning (Dem-AI) lays out a holistic philosophy with underlying principles for building large-scale distributed and democratized machine learning systems.
Inspired by Dem-AI philosophy, a novel distributed learning approach is proposed in this paper.
The proposed algorithms demonstrate better results in the generalization performance of learning models in agents compared to the conventional FL algorithms.
arXiv Detail & Related papers (2020-07-07T08:34:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.