On the (In)Feasibility of Attribute Inference Attacks on Machine
Learning Models
- URL: http://arxiv.org/abs/2103.07101v1
- Date: Fri, 12 Mar 2021 06:21:56 GMT
- Title: On the (In)Feasibility of Attribute Inference Attacks on Machine
Learning Models
- Authors: Benjamin Zi Hao Zhao, Aviral Agrawal, Catisha Coburn, Hassan Jameel
Asghar, Raghav Bhaskar, Mohamed Ali Kaafar, Darren Webb, and Peter Dickinson
- Abstract summary: We show that even if a classification model succumbs to membership inference attacks, it is unlikely to be susceptible to attribute inference attacks.
We show that membership inference attacks cannot infer membership in this strong setting.
Under a relaxed notion of attribute inference, we show that it is possible to infer attributes close to the true attributes.
- Score: 4.1245935888536325
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With an increase in low-cost machine learning APIs, advanced machine learning
models may be trained on private datasets and monetized by providing them as a
service. However, privacy researchers have demonstrated that these models may
leak information about records in the training dataset via membership inference
attacks. In this paper, we take a closer look at another inference attack
reported in literature, called attribute inference, whereby an attacker tries
to infer missing attributes of a partially known record used in the training
dataset by accessing the machine learning model as an API. We show that even if
a classification model succumbs to membership inference attacks, it is unlikely
to be susceptible to attribute inference attacks. We demonstrate that this is
because membership inference attacks fail to distinguish a member from a nearby
non-member. We call the ability of an attacker to distinguish the two (similar)
vectors as strong membership inference. We show that membership inference
attacks cannot infer membership in this strong setting, and hence inferring
attributes is infeasible. However, under a relaxed notion of attribute
inference, called approximate attribute inference, we show that it is possible
to infer attributes close to the true attributes. We verify our results on
three publicly available datasets, five membership, and three attribute
inference attacks reported in literature.
Related papers
- FullCert: Deterministic End-to-End Certification for Training and Inference of Neural Networks [62.897993591443594]
FullCert is the first end-to-end certifier with sound, deterministic bounds.
We experimentally demonstrate FullCert's feasibility on two datasets.
arXiv Detail & Related papers (2024-06-17T13:23:52Z) - Can Adversarial Examples Be Parsed to Reveal Victim Model Information? [62.814751479749695]
In this work, we ask whether it is possible to infer data-agnostic victim model (VM) information from data-specific adversarial instances.
We collect a dataset of adversarial attacks across 7 attack types generated from 135 victim models.
We show that a simple, supervised model parsing network (MPN) is able to infer VM attributes from unseen adversarial attacks.
arXiv Detail & Related papers (2023-03-13T21:21:49Z) - Are Attribute Inference Attacks Just Imputation? [12.56413718364189]
In an attribute inference attack, an adversary has partial knowledge of some training records and access to a model trained on those records.
We show that proposed defenses such as differentially private training and removing vulnerable records from training do not mitigate this privacy risk.
arXiv Detail & Related papers (2022-09-02T23:13:36Z) - Truth Serum: Poisoning Machine Learning Models to Reveal Their Secrets [53.866927712193416]
We show that an adversary who can poison a training dataset can cause models trained on this dataset to leak private details belonging to other parties.
Our attacks are effective across membership inference, attribute inference, and data extraction.
Our results cast doubts on the relevance of cryptographic privacy guarantees in multiparty protocols for machine learning.
arXiv Detail & Related papers (2022-03-31T18:06:28Z) - Are Your Sensitive Attributes Private? Novel Model Inversion Attribute
Inference Attacks on Classification Models [22.569705869469814]
We focus on model inversion attacks where the adversary knows non-sensitive attributes about records in the training data.
We devise a novel confidence score-based model inversion attribute inference attack that significantly outperforms the state-of-the-art.
We also extend our attacks to the scenario where some of the other (non-sensitive) attributes of a target record are unknown to the adversary.
arXiv Detail & Related papers (2022-01-23T21:27:20Z) - Attribute Inference Attack of Speech Emotion Recognition in Federated
Learning Settings [56.93025161787725]
Federated learning (FL) is a distributed machine learning paradigm that coordinates clients to train a model collaboratively without sharing local data.
We propose an attribute inference attack framework that infers sensitive attribute information of the clients from shared gradients or model parameters.
We show that the attribute inference attack is achievable for SER systems trained using FL.
arXiv Detail & Related papers (2021-12-26T16:50:42Z) - Formalizing and Estimating Distribution Inference Risks [11.650381752104298]
We propose a formal and general definition of property inference attacks.
Our results show that inexpensive attacks are as effective as expensive meta-classifier attacks.
We extend the state-of-the-art property inference attack to work on convolutional neural networks.
arXiv Detail & Related papers (2021-09-13T14:54:39Z) - Black-box Model Inversion Attribute Inference Attacks on Classification
Models [32.757792981935815]
We focus on one kind of model inversion attacks, where the adversary knows non-sensitive attributes about instances in the training data.
We devise two novel model inversion attribute inference attacks -- confidence modeling-based attack and confidence score-based attack.
We evaluate our attacks on two types of machine learning models, decision tree and deep neural network, trained with two real datasets.
arXiv Detail & Related papers (2020-12-07T01:14:19Z) - Sampling Attacks: Amplification of Membership Inference Attacks by
Repeated Queries [74.59376038272661]
We introduce sampling attack, a novel membership inference technique that unlike other standard membership adversaries is able to work under severe restriction of no access to scores of the victim model.
We show that a victim model that only publishes the labels is still susceptible to sampling attacks and the adversary can recover up to 100% of its performance.
For defense, we choose differential privacy in the form of gradient perturbation during the training of the victim model as well as output perturbation at prediction time.
arXiv Detail & Related papers (2020-09-01T12:54:54Z) - Label-Only Membership Inference Attacks [67.46072950620247]
We introduce label-only membership inference attacks.
Our attacks evaluate the robustness of a model's predicted labels under perturbations.
We find that training models with differential privacy and (strong) L2 regularization are the only known defense strategies.
arXiv Detail & Related papers (2020-07-28T15:44:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.