Black-Box Ripper: Copying black-box models using generative evolutionary
algorithms
- URL: http://arxiv.org/abs/2010.11158v1
- Date: Wed, 21 Oct 2020 17:25:23 GMT
- Title: Black-Box Ripper: Copying black-box models using generative evolutionary
algorithms
- Authors: Antonio Barbalau, Adrian Cosma, Radu Tudor Ionescu, Marius Popescu
- Abstract summary: We study the task of replicating the functionality of black-box neural models.
We assume back-propagation through the black-box model is not possible.
We present a teacher-student framework that can distill the black-box (teacher) model into a student model.
- Score: 29.243901669124515
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study the task of replicating the functionality of black-box neural
models, for which we only know the output class probabilities provided for a
set of input images. We assume back-propagation through the black-box model is
not possible and its training images are not available, e.g. the model could be
exposed only through an API. In this context, we present a teacher-student
framework that can distill the black-box (teacher) model into a student model
with minimal accuracy loss. To generate useful data samples for training the
student, our framework (i) learns to generate images on a proxy data set (with
images and classes different from those used to train the black-box) and (ii)
applies an evolutionary strategy to make sure that each generated data sample
exhibits a high response for a specific class when given as input to the black
box. Our framework is compared with several baseline and state-of-the-art
methods on three benchmark data sets. The empirical evidence indicates that our
model is superior to the considered baselines. Although our method does not
back-propagate through the black-box network, it generally surpasses
state-of-the-art methods that regard the teacher as a glass-box model. Our code
is available at: https://github.com/antoniobarbalau/black-box-ripper.
Related papers
- Black-Box Forgetting [8.84485103053191]
We address a novel problem of selective forgetting for black-box models, named Black-Box Forgetting.
We propose Latent Context Sharing, which introduces common low-dimensional latent components among multiple tokens for the prompt.
Experiments on four standard benchmark datasets demonstrate the superiority of our method with reasonable baselines.
arXiv Detail & Related papers (2024-11-01T07:10:40Z) - FreeSeg-Diff: Training-Free Open-Vocabulary Segmentation with Diffusion Models [56.71672127740099]
We focus on the task of image segmentation, which is traditionally solved by training models on closed-vocabulary datasets.
We leverage different and relatively small-sized, open-source foundation models for zero-shot open-vocabulary segmentation.
Our approach (dubbed FreeSeg-Diff), which does not rely on any training, outperforms many training-based approaches on both Pascal VOC and COCO datasets.
arXiv Detail & Related papers (2024-03-29T10:38:25Z) - Towards Few-Call Model Stealing via Active Self-Paced Knowledge
Distillation and Diffusion-Based Image Generation [33.60710287553274]
We propose to copy black-box classification models without having access to the original training data, the architecture, and the weights of the model.
We employ a novel active self-paced learning framework to make the most of the proxy data during distillation.
Our empirical results on two data sets confirm the superiority of our framework over two state-of-the-art methods in the few-call model extraction scenario.
arXiv Detail & Related papers (2023-09-29T19:09:27Z) - DREAM: Domain-free Reverse Engineering Attributes of Black-box Model [51.37041886352823]
We propose a new problem of Domain-agnostic Reverse Engineering the Attributes of a black-box target model.
We learn a domain-agnostic model to infer the attributes of a target black-box model with unknown training data.
arXiv Detail & Related papers (2023-07-20T16:25:58Z) - Black-Box Batch Active Learning for Regression [1.52292571922932]
Batch active learning is a popular approach for efficiently training machine learning models on unlabelled datasets.
We propose black-box batch active learning for regression tasks as an extension of white-box approaches.
arXiv Detail & Related papers (2023-02-17T16:35:47Z) - Black-box Few-shot Knowledge Distillation [55.27881513982002]
Knowledge distillation (KD) is an efficient approach to transfer the knowledge from a large "teacher" network to a smaller "student" network.
We propose a black-box few-shot KD method to train the student with few unlabeled training samples and a black-box teacher.
We conduct extensive experiments to show that our method significantly outperforms recent SOTA few/zero-shot KD methods on image classification tasks.
arXiv Detail & Related papers (2022-07-25T12:16:53Z) - How to Robustify Black-Box ML Models? A Zeroth-Order Optimization
Perspective [74.47093382436823]
We address the problem of black-box defense: How to robustify a black-box model using just input queries and output feedback?
We propose a general notion of defensive operation that can be applied to black-box models, and design it through the lens of denoised smoothing (DS)
We empirically show that ZO-AE-DS can achieve improved accuracy, certified robustness, and query complexity over existing baselines.
arXiv Detail & Related papers (2022-03-27T03:23:32Z) - Defending against Model Stealing via Verifying Embedded External
Features [90.29429679125508]
adversaries can steal' deployed models even when they have no training samples and can not get access to the model parameters or structures.
We explore the defense from another angle by verifying whether a suspicious model contains the knowledge of defender-specified emphexternal features.
Our method is effective in detecting different types of model stealing simultaneously, even if the stolen model is obtained via a multi-stage stealing process.
arXiv Detail & Related papers (2021-12-07T03:51:54Z) - Can Explanations Be Useful for Calibrating Black Box Models? [31.473798197405948]
We study how to improve a black box model's performance on a new domain given examples from the new domain.
Our approach first extracts a set of features combining human intuition about the task with model attributions.
We show that the calibration features transfer to some extent between tasks and shed light on how to effectively use them.
arXiv Detail & Related papers (2021-10-14T17:48:16Z) - Visualising Deep Network's Time-Series Representations [93.73198973454944]
Despite the popularisation of machine learning models, more often than not they still operate as black boxes with no insight into what is happening inside the model.
In this paper, a method that addresses that issue is proposed, with a focus on visualising multi-dimensional time-series data.
Experiments on a high-frequency stock market dataset show that the method provides fast and discernible visualisations.
arXiv Detail & Related papers (2021-03-12T09:53:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.