Randomness Concerns When Deploying Differential Privacy
- URL: http://arxiv.org/abs/2009.03777v1
- Date: Sun, 6 Sep 2020 15:28:40 GMT
- Title: Randomness Concerns When Deploying Differential Privacy
- Authors: Simson L. Garfinkel and Philip Leclerc
- Abstract summary: The U.S. Census Bureau is using differential privacy to protect confidential respondent data collected for the 2020 Decennial Census of Population & Housing.
The Census Bureau's DP system is implemented in the Disclosure Avoidance System (DAS) and requires a source of random numbers.
We estimate that the 2020 Census will require roughly 90TB of random bytes to protect the person and household tables.
- Score: 0.25889737226898435
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: The U.S. Census Bureau is using differential privacy (DP) to protect
confidential respondent data collected for the 2020 Decennial Census of
Population & Housing. The Census Bureau's DP system is implemented in the
Disclosure Avoidance System (DAS) and requires a source of random numbers. We
estimate that the 2020 Census will require roughly 90TB of random bytes to
protect the person and household tables. Although there are critical
differences between cryptography and DP, they have similar requirements for
randomness. We review the history of random number generation on deterministic
computers, including von Neumann's "middle-square" method, Mersenne Twister
(MT19937) (previously the default NumPy random number generator, which we
conclude is unacceptable for use in production privacy-preserving systems), and
the Linux /dev/urandom device. We also review hardware random number generator
schemes, including the use of so-called "Lava Lamps" and the Intel Secure Key
RDRAND instruction. We finally present our plan for generating random bits in
the Amazon Web Services (AWS) environment using AES-CTR-DRBG seeded by mixing
bits from /dev/urandom and the Intel Secure Key RDSEED instruction, a
compromise of our desire to rely on a trusted hardware implementation, the
unease of our external reviewers in trusting a hardware-only implementation,
and the need to generate so many random bits.
Related papers
- Data Analysis Methods Preliminaries for a Photon-based Hardware Random Number Generator [39.58317527488534]
Minecraft is great for games, but not so great for cybersecurity encryption key generation.
PRNGs generate random numbers by quantifying some unpredictable phenomena into a number or string and feeding it into an algorithm.
By using a hardware random number generator (HRNG), random numbers that are not susceptible to the flaws found in PRNGs can be attained at a high rate.
arXiv Detail & Related papers (2024-04-15T00:47:17Z) - New Class of Ciphers Using Hardware Entropy Source [0.0]
A stream of random bits is produced by extracting the entropy of a physical process.
The process of placing bits of a message into the stream of random bits is governed by the number of random bits skipped between subsequent insertions.
We propose an effective method of computing random keys from a given number of random bits.
arXiv Detail & Related papers (2024-04-14T15:44:50Z) - A Game-theoretic Approach for Provably-Uniform Random Number Generation in Decentralized Networks [0.6216023343793144]
We provide a protocol for distributed generation of randomness.
It is trustless and generates unbiased random numbers.
It is also tamper-proof and no party can change the output or affect its distribution.
arXiv Detail & Related papers (2023-09-20T12:21:39Z) - Random Boxes Are Open-world Object Detectors [71.86454597677387]
We show that classifiers trained with random region proposals achieve state-of-the-art Open-world Object Detection (OWOD)
We propose RandBox, a Fast R-CNN based architecture trained on random proposals at each training.
RandBox significantly outperforms the previous state-of-the-art in all metrics.
arXiv Detail & Related papers (2023-07-17T05:08:32Z) - Machine Learning needs Better Randomness Standards: Randomised Smoothing
and PRNG-based attacks [14.496582479888765]
We consider whether attackers can compromise an machine learning system using only the randomness on which they commonly rely.
We demonstrate an entirely novel attack, where an attacker backdoors the supplied randomness to falsely certify either an overestimate or an underestimate of robustness for up to 81 times.
We advocate updating the NIST guidelines on random number testing to make them more appropriate for safety-critical and security-critical machine-learning applications.
arXiv Detail & Related papers (2023-06-24T19:50:08Z) - A privacy-preserving publicly verifiable quantum random number generator [48.7576911714538]
We report the implementation of an entanglement-based protocol that allows a third party to publicly perform statistical tests without compromising the privacy of the random bits.
limitations on computing power can restrict an end-user's ability to perform such verification.
arXiv Detail & Related papers (2023-05-18T12:13:48Z) - Testing randomness of series generated in Bell's experiment [62.997667081978825]
We use a toy fiber optic based setup to generate binary series, and evaluate their level of randomness according to Ville principle.
Series are tested with a battery of standard statistical indicators, Hurst, Kolmogorov complexity, minimum entropy, Takensarity dimension of embedding, and Augmented Dickey Fuller and Kwiatkowski Phillips Schmidt Shin to check station exponent.
The level of randomness of series obtained by applying Toeplitz extractor to rejected series is found to be indistinguishable from the level of non-rejected raw ones.
arXiv Detail & Related papers (2022-08-31T17:39:29Z) - Syfer: Neural Obfuscation for Private Data Release [58.490998583666276]
We develop Syfer, a neural obfuscation method to protect against re-identification attacks.
Syfer composes trained layers with random neural networks to encode the original data.
It maintains the ability to predict diagnoses from the encoded data.
arXiv Detail & Related papers (2022-01-28T20:32:04Z) - Certified Random Number Generation from Quantum Steering [1.0820909926464386]
Certified randomness protocols have been developed which remove the need for trust in devices by taking advantage of nonlocality.
Here, we use a photonic platform to implement our protocol, which operates in the quantum steering scenario.
We demonstrate an approach for a steering-based generator of public or private randomness, and the first generation of certified random bits, with the detection loophole closed.
arXiv Detail & Related papers (2021-11-18T03:49:43Z) - NeuraCrypt: Hiding Private Health Data via Random Neural Networks for
Public Training [64.54200987493573]
We propose NeuraCrypt, a private encoding scheme based on random deep neural networks.
NeuraCrypt encodes raw patient data using a randomly constructed neural network known only to the data-owner.
We show that NeuraCrypt achieves competitive accuracy to non-private baselines on a variety of x-ray tasks.
arXiv Detail & Related papers (2021-06-04T13:42:21Z) - Lossless Compression of Efficient Private Local Randomizers [55.657133416044104]
Locally Differentially Private (LDP) Reports are commonly used for collection of statistics and machine learning in the federated setting.
In many cases the best known LDP algorithms require sending prohibitively large messages from the client device to the server.
This has led to significant efforts on reducing the communication cost of LDP algorithms.
arXiv Detail & Related papers (2021-02-24T07:04:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.