DREAM: Domain-free Reverse Engineering Attributes of Black-box Model
- URL: http://arxiv.org/abs/2307.10997v1
- Date: Thu, 20 Jul 2023 16:25:58 GMT
- Title: DREAM: Domain-free Reverse Engineering Attributes of Black-box Model
- Authors: Rongqing Li, Jiaqi Yu, Changsheng Li, Wenhan Luo, Ye Yuan, Guoren Wang
- Abstract summary: We propose a new problem of Domain-agnostic Reverse Engineering the Attributes of a black-box target model.
We learn a domain-agnostic model to infer the attributes of a target black-box model with unknown training data.
- Score: 51.37041886352823
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning models are usually black boxes when deployed on machine
learning platforms. Prior works have shown that the attributes ($e.g.$, the
number of convolutional layers) of a target black-box neural network can be
exposed through a sequence of queries. There is a crucial limitation: these
works assume the dataset used for training the target model to be known
beforehand and leverage this dataset for model attribute attack. However, it is
difficult to access the training dataset of the target black-box model in
reality. Therefore, whether the attributes of a target black-box model could be
still revealed in this case is doubtful. In this paper, we investigate a new
problem of Domain-agnostic Reverse Engineering the Attributes of a black-box
target Model, called DREAM, without requiring the availability of the target
model's training dataset, and put forward a general and principled framework by
casting this problem as an out of distribution (OOD) generalization problem. In
this way, we can learn a domain-agnostic model to inversely infer the
attributes of a target black-box model with unknown training data. This makes
our method one of the kinds that can gracefully apply to an arbitrary domain
for model attribute reverse engineering with strong generalization ability.
Extensive experimental studies are conducted and the results validate the
superiority of our proposed method over the baselines.
Related papers
- Cross-Domain Transfer Learning with CoRTe: Consistent and Reliable
Transfer from Black-Box to Lightweight Segmentation Model [25.3403116022412]
CoRTe is a pseudo-labelling function that extracts reliable knowledge from a black-box source model.
We benchmark CoRTe on two synthetic-to-real settings, demonstrating remarkable results when using black-box models to transfer knowledge on lightweight models for a target data distribution.
arXiv Detail & Related papers (2024-02-20T16:35:14Z) - Data-Free Model Extraction Attacks in the Context of Object Detection [0.6719751155411076]
A significant number of machine learning models are vulnerable to model extraction attacks.
We propose an adversary black box attack extending to a regression problem for predicting bounding box coordinates in object detection.
We find that the proposed model extraction method achieves significant results by using reasonable queries.
arXiv Detail & Related papers (2023-08-09T06:23:54Z) - Distilling BlackBox to Interpretable models for Efficient Transfer
Learning [19.40897632956169]
Building generalizable AI models is one of the primary challenges in the healthcare domain.
Fine-tuning a model to transfer knowledge from one domain to another requires a significant amount of labeled data in the target domain.
We develop an interpretable model that can be efficiently fine-tuned to an unseen target domain with minimal computational cost.
arXiv Detail & Related papers (2023-05-26T23:23:48Z) - Query Efficient Cross-Dataset Transferable Black-Box Attack on Action
Recognition [99.29804193431823]
Black-box adversarial attacks present a realistic threat to action recognition systems.
We propose a new attack on action recognition that addresses these shortcomings by generating perturbations.
Our method achieves 8% and higher 12% deception rates compared to state-of-the-art query-based and transfer-based attacks.
arXiv Detail & Related papers (2022-11-23T17:47:49Z) - RAIN: RegulArization on Input and Network for Black-Box Domain
Adaptation [80.03883315743715]
Source-free domain adaptation transits the source-trained model towards target domain without exposing the source data.
This paradigm is still at risk of data leakage due to adversarial attacks on the source model.
We propose a novel approach named RAIN (RegulArization on Input and Network) for Black-Box domain adaptation from both input-level and network-level regularization.
arXiv Detail & Related papers (2022-08-22T18:18:47Z) - How to Robustify Black-Box ML Models? A Zeroth-Order Optimization
Perspective [74.47093382436823]
We address the problem of black-box defense: How to robustify a black-box model using just input queries and output feedback?
We propose a general notion of defensive operation that can be applied to black-box models, and design it through the lens of denoised smoothing (DS)
We empirically show that ZO-AE-DS can achieve improved accuracy, certified robustness, and query complexity over existing baselines.
arXiv Detail & Related papers (2022-03-27T03:23:32Z) - Label-Only Model Inversion Attacks via Boundary Repulsion [12.374249336222906]
We introduce an algorithm to invert private training data using only the target model's predicted labels.
Using the example of face recognition, we show that the images reconstructed by BREP-MI successfully reproduce the semantics of the private training data.
arXiv Detail & Related papers (2022-03-03T18:57:57Z) - Distill and Fine-tune: Effective Adaptation from a Black-box Source
Model [138.12678159620248]
Unsupervised domain adaptation (UDA) aims to transfer knowledge in previous related labeled datasets (source) to a new unlabeled dataset (target)
We propose a novel two-step adaptation framework called Distill and Fine-tune (Dis-tune)
arXiv Detail & Related papers (2021-04-04T05:29:05Z) - Design of Dynamic Experiments for Black-Box Model Discrimination [72.2414939419588]
Consider a dynamic model discrimination setting where we wish to chose: (i) what is the best mechanistic, time-varying model and (ii) what are the best model parameter estimates.
For rival mechanistic models where we have access to gradient information, we extend existing methods to incorporate a wider range of problem uncertainty.
We replace these black-box models with Gaussian process surrogate models and thereby extend the model discrimination setting to additionally incorporate rival black-box model.
arXiv Detail & Related papers (2021-02-07T11:34:39Z) - REST: Performance Improvement of a Black Box Model via RL-based Spatial
Transformation [15.691668909002892]
We study robustness to geometric transformations in a specific condition where the black-box image classifier is given.
We propose an additional learner, emphREinforcement Spatial Transform (REST), that transforms the warped input data into samples regarded as in-distribution by the black-box models.
arXiv Detail & Related papers (2020-02-16T16:15:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.