Learning to Learn from APIs: Black-Box Data-Free Meta-Learning
- URL: http://arxiv.org/abs/2305.18413v2
- Date: Mon, 19 Jun 2023 15:25:37 GMT
- Title: Learning to Learn from APIs: Black-Box Data-Free Meta-Learning
- Authors: Zixuan Hu, Li Shen, Zhenyi Wang, Baoyuan Wu, Chun Yuan, Dacheng Tao
- Abstract summary: Data-free meta-learning (DFML) aims to enable efficient learning of new tasks by meta-learning from a collection of pre-trained models without access to the training data.
Existing DFML work can only meta-learn from (i) white-box and (ii) small-scale pre-trained models.
We propose a Bi-level Data-free Meta Knowledge Distillation (BiDf-MKD) framework to transfer more general meta knowledge from a collection of black-box APIs to one single model.
- Score: 95.41441357931397
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Data-free meta-learning (DFML) aims to enable efficient learning of new tasks
by meta-learning from a collection of pre-trained models without access to the
training data. Existing DFML work can only meta-learn from (i) white-box and
(ii) small-scale pre-trained models (iii) with the same architecture,
neglecting the more practical setting where the users only have inference
access to the APIs with arbitrary model architectures and model scale inside.
To solve this issue, we propose a Bi-level Data-free Meta Knowledge
Distillation (BiDf-MKD) framework to transfer more general meta knowledge from
a collection of black-box APIs to one single meta model. Specifically, by just
querying APIs, we inverse each API to recover its training data via a
zero-order gradient estimator and then perform meta-learning via a novel
bi-level meta knowledge distillation structure, in which we design a boundary
query set recovery technique to recover a more informative query set near the
decision boundary. In addition, to encourage better generalization within the
setting of limited API budgets, we propose task memory replay to diversify the
underlying task distribution by covering more interpolated tasks. Extensive
experiments in various real-world scenarios show the superior performance of
our BiDf-MKD framework.
Related papers
- Privacy Challenges in Meta-Learning: An Investigation on Model-Agnostic Meta-Learning [2.8948274245812327]
In current meta-learning methods, task learners locally learn models from sensitive data, termed support sets.
Despite the absence of explicit data sharing, privacy concerns persist.
This paper examines potential data leakage in a prominent metalearning algorithm, specifically Model-Agnostic Meta-Learning (MAML)
arXiv Detail & Related papers (2024-06-01T01:10:35Z) - FREE: Faster and Better Data-Free Meta-Learning [77.90126669914324]
Data-Free Meta-Learning (DFML) aims to extract knowledge from a collection of pre-trained models without requiring the original data.
We introduce the Faster and Better Data-Free Meta-Learning framework, which contains: (i) a meta-generator for rapidly recovering training tasks from pre-trained models; and (ii) a meta-learner for generalizing to new unseen tasks.
arXiv Detail & Related papers (2024-05-02T03:43:19Z) - API-BLEND: A Comprehensive Corpora for Training and Benchmarking API LLMs [28.840207102132286]
We focus on the task of identifying, curating, and transforming existing datasets.
We introduce API-BLEND, a large corpora for training and systematic testing of tool-augmented LLMs.
We demonstrate the utility of the API-BLEND dataset for both training and benchmarking purposes.
arXiv Detail & Related papers (2024-02-23T18:30:49Z) - Architecture, Dataset and Model-Scale Agnostic Data-free Meta-Learning [119.70303730341938]
We propose ePisode cUrriculum inveRsion (ECI) during data-free meta training and invErsion calibRation following inner loop (ICFIL) during meta testing.
ECI adaptively increases the difficulty level of pseudo episodes according to the real-time feedback of the meta model.
We formulate the optimization process of meta training with ECI as an adversarial form in an end-to-end manner.
arXiv Detail & Related papers (2023-03-20T15:10:41Z) - Meta-Regularization by Enforcing Mutual-Exclusiveness [0.8057006406834467]
We propose a regularization technique for meta-learning models that gives the model designer more control over the information flow during meta-training.
Our proposed regularization function shows an accuracy boost of $sim$ $36%$ on the Omniglot dataset.
arXiv Detail & Related papers (2021-01-24T22:57:19Z) - Learning to Generalize Unseen Domains via Memory-based Multi-Source
Meta-Learning for Person Re-Identification [59.326456778057384]
We propose the Memory-based Multi-Source Meta-Learning framework to train a generalizable model for unseen domains.
We also present a meta batch normalization layer (MetaBN) to diversify meta-test features.
Experiments demonstrate that our M$3$L can effectively enhance the generalization ability of the model for unseen domains.
arXiv Detail & Related papers (2020-12-01T11:38:16Z) - Fast Few-Shot Classification by Few-Iteration Meta-Learning [173.32497326674775]
We introduce a fast optimization-based meta-learning method for few-shot classification.
Our strategy enables important aspects of the base learner objective to be learned during meta-training.
We perform a comprehensive experimental analysis, demonstrating the speed and effectiveness of our approach.
arXiv Detail & Related papers (2020-10-01T15:59:31Z) - Incremental Meta-Learning via Indirect Discriminant Alignment [118.61152684795178]
We develop a notion of incremental learning during the meta-training phase of meta-learning.
Our approach performs favorably at test time as compared to training a model with the full meta-training set.
arXiv Detail & Related papers (2020-02-11T01:39:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.