Estimating g-Leakage via Machine Learning
- URL: http://arxiv.org/abs/2005.04399v3
- Date: Wed, 24 Nov 2021 22:05:32 GMT
- Title: Estimating g-Leakage via Machine Learning
- Authors: Marco Romanelli and Konstantinos Chatzikokolakis and Catuscia
Palamidessi and Pablo Piantanida
- Abstract summary: This paper considers the problem of estimating the information leakage of a system in the black-box scenario.
It is assumed that the system's internals are unknown to the learner, or anyway too complicated to analyze.
We propose a novel approach to perform black-box estimation of the g-vulnerability using Machine Learning (ML) algorithms.
- Score: 34.102705643128004
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper considers the problem of estimating the information leakage of a
system in the black-box scenario. It is assumed that the system's internals are
unknown to the learner, or anyway too complicated to analyze, and the only
available information are pairs of input-output data samples, possibly obtained
by submitting queries to the system or provided by a third party. Previous
research has mainly focused on counting the frequencies to estimate the
input-output conditional probabilities (referred to as frequentist approach),
however this method is not accurate when the domain of possible outputs is
large. To overcome this difficulty, the estimation of the Bayes error of the
ideal classifier was recently investigated using Machine Learning (ML) models
and it has been shown to be more accurate thanks to the ability of those models
to learn the input-output correspondence. However, the Bayes vulnerability is
only suitable to describe one-try attacks. A more general and flexible measure
of leakage is the g-vulnerability, which encompasses several different types of
adversaries, with different goals and capabilities. In this paper, we propose a
novel approach to perform black-box estimation of the g-vulnerability using ML.
A feature of our approach is that it does not require to estimate the
conditional probabilities, and that it is suitable for a large class of ML
algorithms. First, we formally show the learnability for all data
distributions. Then, we evaluate the performance via various experiments using
k-Nearest Neighbors and Neural Networks. Our results outperform the frequentist
approach when the observables domain is large.
Related papers
- Estimating Uncertainty with Implicit Quantile Network [0.0]
Uncertainty quantification is an important part of many performance critical applications.
This paper provides a simple alternative to existing approaches such as ensemble learning and bayesian neural networks.
arXiv Detail & Related papers (2024-08-26T13:33:14Z) - Probabilistic Contrastive Learning for Long-Tailed Visual Recognition [78.70453964041718]
Longtailed distributions frequently emerge in real-world data, where a large number of minority categories contain a limited number of samples.
Recent investigations have revealed that supervised contrastive learning exhibits promising potential in alleviating the data imbalance.
We propose a novel probabilistic contrastive (ProCo) learning algorithm that estimates the data distribution of the samples from each class in the feature space.
arXiv Detail & Related papers (2024-03-11T13:44:49Z) - Do Bayesian Variational Autoencoders Know What They Don't Know? [0.6091702876917279]
The problem of detecting the Out-of-Distribution (OoD) inputs is paramount importance for Deep Neural Networks.
It has been previously shown that even Deep Generative Models that allow estimating the density of the inputs may not be reliable.
This paper investigates three approaches to inference: Markov chain Monte Carlo, Bayes gradient by Backpropagation and Weight Averaging-Gaussian.
arXiv Detail & Related papers (2022-12-29T11:48:01Z) - Transformers Can Do Bayesian Inference [56.99390658880008]
We present Prior-Data Fitted Networks (PFNs)
PFNs leverage in-context learning in large-scale machine learning techniques to approximate a large set of posteriors.
We demonstrate that PFNs can near-perfectly mimic Gaussian processes and also enable efficient Bayesian inference for intractable problems.
arXiv Detail & Related papers (2021-12-20T13:07:39Z) - The information of attribute uncertainties: what convolutional neural
networks can learn about errors in input data [0.0]
We show how Convolutional Neural Networks (CNNs) are able to learn about the context and patterns of signal and noise.
We show that, when each data point is subject to different levels of noise, that information can be learned by the CNNs.
arXiv Detail & Related papers (2021-08-10T15:10:46Z) - A Gentle Introduction to Conformal Prediction and Distribution-Free
Uncertainty Quantification [1.90365714903665]
This hands-on introduction is aimed at a reader interested in the practical implementation of distribution-free UQ.
We will include many explanatory illustrations, examples, and code samples in Python, with PyTorch syntax.
arXiv Detail & Related papers (2021-07-15T17:59:50Z) - Evaluating State-of-the-Art Classification Models Against Bayes
Optimality [106.50867011164584]
We show that we can compute the exact Bayes error of generative models learned using normalizing flows.
We use our approach to conduct a thorough investigation of state-of-the-art classification models.
arXiv Detail & Related papers (2021-06-07T06:21:20Z) - DAAIN: Detection of Anomalous and Adversarial Input using Normalizing
Flows [52.31831255787147]
We introduce a novel technique, DAAIN, to detect out-of-distribution (OOD) inputs and adversarial attacks (AA)
Our approach monitors the inner workings of a neural network and learns a density estimator of the activation distribution.
Our model can be trained on a single GPU making it compute efficient and deployable without requiring specialized accelerators.
arXiv Detail & Related papers (2021-05-30T22:07:13Z) - Provable tradeoffs in adversarially robust classification [96.48180210364893]
We develop and leverage new tools, including recent breakthroughs from probability theory on robust isoperimetry.
Our results reveal fundamental tradeoffs between standard and robust accuracy that grow when data is imbalanced.
arXiv Detail & Related papers (2020-06-09T09:58:19Z) - Meta-Learned Confidence for Few-shot Learning [60.6086305523402]
A popular transductive inference technique for few-shot metric-based approaches, is to update the prototype of each class with the mean of the most confident query examples.
We propose to meta-learn the confidence for each query sample, to assign optimal weights to unlabeled queries.
We validate our few-shot learning model with meta-learned confidence on four benchmark datasets.
arXiv Detail & Related papers (2020-02-27T10:22:17Z) - Multi-class Gaussian Process Classification with Noisy Inputs [2.362412515574206]
In some situations, the amount of noise can be known before-hand.
We have evaluated the proposed methods by carrying out several experiments, involving synthetic and real data.
arXiv Detail & Related papers (2020-01-28T18:55:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.