Black-box Prompt Learning for Pre-trained Language Models
- URL: http://arxiv.org/abs/2201.08531v1
- Date: Fri, 21 Jan 2022 03:53:19 GMT
- Title: Black-box Prompt Learning for Pre-trained Language Models
- Authors: Shizhe Diao, Xuechun Li, Yong Lin, Zhichao Huang, Tong Zhang
- Abstract summary: This work considers a new scenario, where we do not have access to a pre-trained model, except for its outputs given inputs.
We first introduce the black-box setting formally on text classification, where the pre-trained model is not only frozen but also invisible.
We then propose our solution black-box prompt, a new technique in the prompt-learning family, which can leverage the knowledge learned by pre-trained models from the pre-training corpus.
- Score: 18.17029934303874
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Domain-specific fine-tuning strategies for large pre-trained models received
vast attention in recent years. In previously studied settings, the model
architectures and parameters are tunable or at least visible, which we refer to
as white-box settings. This work considers a new scenario, where we do not have
access to a pre-trained model, except for its outputs given inputs, and we call
this problem black-box fine-tuning. To illustrate our approach, we first
introduce the black-box setting formally on text classification, where the
pre-trained model is not only frozen but also invisible. We then propose our
solution black-box prompt, a new technique in the prompt-learning family, which
can leverage the knowledge learned by pre-trained models from the pre-training
corpus. Our experiments demonstrate that the proposed method achieved the
state-of-the-art performance on eight datasets. Further analyses on different
human-designed objectives, prompt lengths, and intuitive explanations
demonstrate the robustness and flexibility of our method.
Related papers
- IOTA: Corrective Knowledge-Guided Prompt Learning via Black-White Box Framework [57.66924056568018]
We propose a novel black-whIte bOx prompT leArning framework (IOTA) for adapting pre-trained models to downstream tasks.<n>IOTA integrates a data-driven Black Box module with a knowledge-driven White Box module for downstream task adaptation.
arXiv Detail & Related papers (2026-01-28T12:03:48Z) - Personalizing black-box models for nonparametric regression with minimax optimality [17.981373446046366]
We study few-shot personalization, where a pre-trained black-box model is adapted to a target domain using a limited number of samples.<n>We propose algorithms that can incorporate a black-box pre-trained model into the regression procedure.
arXiv Detail & Related papers (2026-01-04T08:32:28Z) - On Transfer-based Universal Attacks in Pure Black-box Setting [94.92884394009288]
We study the role of prior knowledge of the target model data and number of classes in attack performance.
We also provide several interesting insights based on our analysis, and demonstrate that priors cause overestimation in transferability scores.
arXiv Detail & Related papers (2025-04-11T10:41:20Z) - Training Spatial-Frequency Visual Prompts and Probabilistic Clusters for Accurate Black-Box Transfer Learning [35.72926400167876]
This paper proposes a novel parameter-efficient transfer learning framework for vision recognition models in the black-box setting.
In experiments, our model demonstrates superior performance in a few-shot transfer learning setting across extensive visual recognition datasets.
arXiv Detail & Related papers (2024-08-15T05:35:52Z) - Parameter-Efficient and Memory-Efficient Tuning for Vision Transformer: A Disentangled Approach [87.8330887605381]
We show how to adapt a pre-trained Vision Transformer to downstream recognition tasks with only a few learnable parameters.
We synthesize a task-specific query with a learnable and lightweight module, which is independent of the pre-trained model.
Our method achieves state-of-the-art performance under memory constraints, showcasing its applicability in real-world situations.
arXiv Detail & Related papers (2024-07-09T15:45:04Z) - Mafin: Enhancing Black-Box Embeddings with Model Augmented Fine-Tuning [13.211063836237468]
We introduce Model augmented fine-tuning (Mafin) -- a novel approach for fine-tuning a black-box embedding model by augmenting it with a trainable embedding model.
Our results demonstrate that Mafin significantly enhances the performance of the black-box embeddings by only requiring the training of a small augmented model.
arXiv Detail & Related papers (2024-02-19T14:33:24Z) - Black-Box Tuning of Vision-Language Models with Effective Gradient
Approximation [71.21346469382821]
We introduce collaborative black-box tuning (CBBT) for both textual prompt optimization and output feature adaptation for black-box models.
CBBT is extensively evaluated on eleven downstream benchmarks and achieves remarkable improvements compared to existing black-box VL adaptation methods.
arXiv Detail & Related papers (2023-12-26T06:31:28Z) - Enhancing Black-Box Few-Shot Text Classification with Prompt-Based Data
Augmentation [42.05617728412819]
We show how to optimize few-shot text classification without accessing the gradients of the large-scale language models.
Our approach, dubbed BT-Classifier, significantly outperforms state-of-the-art black-box few-shot learners.
arXiv Detail & Related papers (2023-05-23T07:54:34Z) - Gradient-Regulated Meta-Prompt Learning for Generalizable
Vision-Language Models [137.74524357614285]
We introduce a novel Gradient-RegulAted Meta-prompt learning framework.
It helps pre-training models adapt to downstream tasks in a parameter -- and data -- efficient way.
GRAM can be easily incorporated into various prompt tuning methods in a model-agnostic way.
arXiv Detail & Related papers (2023-03-12T05:03:37Z) - Self-Distillation for Further Pre-training of Transformers [83.84227016847096]
We propose self-distillation as a regularization for a further pre-training stage.
We empirically validate the efficacy of self-distillation on a variety of benchmark datasets for image and text classification tasks.
arXiv Detail & Related papers (2022-09-30T02:25:12Z) - Reinforcement Learning with Action-Free Pre-Training from Videos [95.25074614579646]
We introduce a framework that learns representations useful for understanding the dynamics via generative pre-training on videos.
Our framework significantly improves both final performances and sample-efficiency of vision-based reinforcement learning.
arXiv Detail & Related papers (2022-03-25T19:44:09Z) - Can Explanations Be Useful for Calibrating Black Box Models? [31.473798197405948]
We study how to improve a black box model's performance on a new domain given examples from the new domain.
Our approach first extracts a set of features combining human intuition about the task with model attributions.
We show that the calibration features transfer to some extent between tasks and shed light on how to effectively use them.
arXiv Detail & Related papers (2021-10-14T17:48:16Z) - Pre-train, Prompt, and Predict: A Systematic Survey of Prompting Methods
in Natural Language Processing [78.8500633981247]
This paper surveys and organizes research works in a new paradigm in natural language processing, which we dub "prompt-based learning"
Unlike traditional supervised learning, which trains a model to take in an input x and predict an output y as P(y|x), prompt-based learning is based on language models that model the probability of text directly.
arXiv Detail & Related papers (2021-07-28T18:09:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.