Towards Sustainable Self-supervised Learning
- URL: http://arxiv.org/abs/2210.11016v1
- Date: Thu, 20 Oct 2022 04:49:56 GMT
- Title: Towards Sustainable Self-supervised Learning
- Authors: Shanghua Gao, Pan Zhou, Ming-Ming Cheng, Shuicheng Yan
- Abstract summary: We propose a Target-Enhanced Conditional (TEC) scheme which introduces two components to the existing mask-reconstruction based SSL.
First, we propose patch-relation enhanced targets which enhances the target given by base model and encourages the new model to learn semantic-relation knowledge from the base model.
Secondly, we introduce a conditional adapter that adaptively adjusts new model prediction to align with the target of different base models.
- Score: 193.78876000005366
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Although increasingly training-expensive, most self-supervised learning (SSL)
models have repeatedly been trained from scratch but not fully utilized, since
only a few SOTAs are employed for downstream tasks. In this work, we explore a
sustainable SSL framework with two major challenges: i) learning a stronger new
SSL model based on the existing pretrained SSL model, also called as "base"
model, in a cost-friendly manner, ii) allowing the training of the new model to
be compatible with various base models. We propose a Target-Enhanced
Conditional (TEC) scheme which introduces two components to the existing
mask-reconstruction based SSL. Firstly, we propose patch-relation enhanced
targets which enhances the target given by base model and encourages the new
model to learn semantic-relation knowledge from the base model by using
incomplete inputs. This hardening and target-enhancing help the new model
surpass the base model, since they enforce additional patch relation modeling
to handle incomplete input. Secondly, we introduce a conditional adapter that
adaptively adjusts new model prediction to align with the target of different
base models. Extensive experimental results show that our TEC scheme can
accelerate the learning speed, and also improve SOTA SSL base models, e.g., MAE
and iBOT, taking an explorative step towards sustainable SSL.
Related papers
- Efficient Training of Self-Supervised Speech Foundation Models on a
Compute Budget [57.807614181024114]
This paper investigates how to efficiently train speech foundation models with self-supervised learning (SSL) under a limited compute budget.
We examine critical factors in SSL that impact the budget, including model architecture, model size, and data size.
arXiv Detail & Related papers (2024-09-09T10:36:42Z) - LLM Augmented LLMs: Expanding Capabilities through Composition [56.40953749310957]
CALM -- Composition to Augment Language Models -- introduces cross-attention between models to compose their representations and enable new capabilities.
We illustrate that augmenting PaLM2-S with a smaller model trained on low-resource languages results in an absolute improvement of up to 13% on tasks like translation into English.
When PaLM2-S is augmented with a code-specific model, we see a relative improvement of 40% over the base model for code generation and explanation tasks.
arXiv Detail & Related papers (2024-01-04T18:53:01Z) - TaCA: Upgrading Your Visual Foundation Model with Task-agnostic
Compatible Adapter [21.41170708560114]
A growing number of applications based on visual foundation models are emerging.
In situations involving system upgrades, it becomes essential to re-train all downstream modules to adapt to the new foundation model.
We introduce a parameter-efficient and task-agnostic adapter, dubbed TaCA, that facilitates compatibility across distinct foundation models.
arXiv Detail & Related papers (2023-06-22T03:00:24Z) - Pushing the Limits of Unsupervised Unit Discovery for SSL Speech
Representation [12.506633315768832]
HuBERT is a successful example that utilizes offline clustering to convert speech features into discrete units for a masked language modeling pretext task.
We present an unsupervised method to improve SSL targets.
Two models are proposed, MonoBERT and PolyBERT, which leverage context-independent and context-dependent phoneme-based units for pre-training.
arXiv Detail & Related papers (2023-06-15T07:45:12Z) - Universal Domain Adaptation from Foundation Models: A Baseline Study [58.51162198585434]
We make empirical studies of state-of-the-art UniDA methods using foundation models.
We introduce textitCLIP distillation, a parameter-free method specifically designed to distill target knowledge from CLIP models.
Although simple, our method outperforms previous approaches in most benchmark tasks.
arXiv Detail & Related papers (2023-05-18T16:28:29Z) - MooseNet: A Trainable Metric for Synthesized Speech with a PLDA Module [3.42658286826597]
We present MooseNet, a trainable speech metric that predicts the listeners' Mean Opinion Score (MOS)
We propose a novel approach where the Probabilistic Linear Discriminative Analysis (PLDA) generative model is used on top of an embedding.
We show that PLDA works well with a non-finetuned SSL model when trained only on 136 utterances.
arXiv Detail & Related papers (2023-01-17T18:53:15Z) - A Strong Baseline for Semi-Supervised Incremental Few-Shot Learning [54.617688468341704]
Few-shot learning aims to learn models that generalize to novel classes with limited training samples.
We propose a novel paradigm containing two parts: (1) a well-designed meta-training algorithm for mitigating ambiguity between base and novel classes caused by unreliable pseudo labels and (2) a model adaptation mechanism to learn discriminative features for novel classes while preserving base knowledge using few labeled and all the unlabeled data.
arXiv Detail & Related papers (2021-10-21T13:25:52Z) - Meta-Learned Attribute Self-Gating for Continual Generalized Zero-Shot
Learning [82.07273754143547]
We propose a meta-continual zero-shot learning (MCZSL) approach to generalizing a model to categories unseen during training.
By pairing self-gating of attributes and scaled class normalization with meta-learning based training, we are able to outperform state-of-the-art results.
arXiv Detail & Related papers (2021-02-23T18:36:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.