Differential Perspectives: Epistemic Disconnects Surrounding the US Census Bureau's Use of Differential Privacy
- URL: http://arxiv.org/abs/2602.18648v1
- Date: Fri, 20 Feb 2026 22:49:07 GMT
- Title: Differential Perspectives: Epistemic Disconnects Surrounding the US Census Bureau's Use of Differential Privacy
- Authors: Danah Boyd, Jayshree Sarathy,
- Abstract summary: The U.S. Census Bureau announced its intention to modernize its disclosure avoidance procedures for the 2020 Census.<n>The move to differential privacy introduced technical and procedural uncertainties, leaving stakeholders unable to evaluate the quality of the data.<n>This essay examines the current controversy over differential privacy as a battle over uncertainty, trust, and legitimacy of the Census.
- Score: 0.9519662954536724
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: When the U.S. Census Bureau announced its intention to modernize its disclosure avoidance procedures for the 2020 Census, it sparked a controversy that is still underway. The move to differential privacy introduced technical and procedural uncertainties, leaving stakeholders unable to evaluate the quality of the data. More importantly, this transformation exposed the statistical illusions and limitations of census data, weakening stakeholders' trust in the data and in the Census Bureau itself. This essay examines the epistemic currents of this controversy. Drawing on theories from Science and Technology Studies (STS) and ethnographic fieldwork, we analyze the current controversy over differential privacy as a battle over uncertainty, trust, and legitimacy of the Census. We argue that rebuilding trust will require more than technical repairs or improved communication; it will require reconstructing what we identify as a 'statistical imaginary.'
Related papers
- Statistical Imaginaries, State Legitimacy: Grappling with the Arrangements Underpinning Quantification in the U.S. Census [0.9519662954536724]
We argue that controversies emerge from $textitarrangements of statistical imaginaries.<n>We compare reactions to two methods designed to improve data accuracy.<n>We show how the credibility of the Census Bureau and its data stem from how statistical imaginaries are contested and stabilized.
arXiv Detail & Related papers (2026-02-20T22:07:56Z) - Partial Identification Approach to Counterfactual Fairness Assessment [50.88100567472179]
We introduce a Bayesian approach to bound unknown counterfactual fairness measures with high confidence.<n>Our results reveal a positive (spurious) effect on the COMPAS score when changing race to African-American (from all others) and a negative (direct causal) effect when transitioning from young to old age.
arXiv Detail & Related papers (2025-09-30T18:35:08Z) - Evaluating the Impacts of Swapping on the US Decennial Census [7.020785266789317]
We describe and implement a parameterized swapping algorithm based on Census publications, court documents, and informal interviews with Census employees.<n>We provide intuition for the types of shifts induced by swapping and compare against those introduced by TopDown.
arXiv Detail & Related papers (2025-02-03T12:51:16Z) - The 2020 United States Decennial Census Is More Private Than You (Might) Think [21.96007825598595]
We show that the 2020 U.S. Census provides significantly stronger privacy protections than its nominal guarantees suggest.<n>We show that noise variances could be reduced by $15.08%$ to $24.82%$ while maintaining nearly the same level of privacy protection for each geographical level.
arXiv Detail & Related papers (2024-10-11T23:06:15Z) - The Complexities of Differential Privacy for Survey Data [0.0]
The U.S. Census Bureau announced the adoption of the concept for its 2020 Decennial Census.
Despite its attractive theoretical properties, implementing DP in practice remains challenging, especially when it comes to survey data.
We identify five aspects that need to be considered when adopting DP in the survey context.
arXiv Detail & Related papers (2024-08-13T16:15:42Z) - Differentially Private Data Release on Graphs: Inefficiencies and Unfairness [48.96399034594329]
This paper characterizes the impact of Differential Privacy on bias and unfairness in the context of releasing information about networks.
We consider a network release problem where the network structure is known to all, but the weights on edges must be released privately.
Our work provides theoretical foundations and empirical evidence into the bias and unfairness arising due to privacy in these networked decision problems.
arXiv Detail & Related papers (2024-08-08T08:37:37Z) - Making Differential Privacy Work for Census Data Users [0.0]
The U.S. Census Bureau collects and publishes detailed demographic data about Americans which are heavily used by researchers and policymakers.
A key output of this privacy protection system is the Noisy Measurement File (NMF), which is produced by adding random noise to tabulated statistics.
We describe the process we use to transform the NMF into a usable format, and provide recommendations to the Bureau for how to release future versions of the NMF.
arXiv Detail & Related papers (2023-05-12T02:48:11Z) - Comment: The Essential Role of Policy Evaluation for the 2020 Census
Disclosure Avoidance System [0.0]
boyd and Sarathy, "Differential Perspectives: Epistemic Disconnects Surrounding the US Census Bureau's Use of Differential Privacy"
We argue that empirical evaluations of the Census Disclosure Avoidance System failed to recognize how the benchmark data is never a ground truth of population counts.
We argue that policy makers must confront a key trade-off between data utility and privacy protection.
arXiv Detail & Related papers (2022-10-15T21:41:54Z) - Post-processing of Differentially Private Data: A Fairness Perspective [53.29035917495491]
This paper shows that post-processing causes disparate impacts on individuals or groups.
It analyzes two critical settings: the release of differentially private datasets and the use of such private datasets for downstream decisions.
It proposes a novel post-processing mechanism that is (approximately) optimal under different fairness metrics.
arXiv Detail & Related papers (2022-01-24T02:45:03Z) - Decision Making with Differential Privacy under a Fairness Lens [65.16089054531395]
The U.S. Census Bureau releases data sets and statistics about groups of individuals that are used as input to a number of critical decision processes.
To conform to privacy and confidentiality requirements, these agencies are often required to release privacy-preserving versions of the data.
This paper studies the release of differentially private data sets and analyzes their impact on some critical resource allocation tasks under a fairness perspective.
arXiv Detail & Related papers (2021-05-16T21:04:19Z) - Robustness Threats of Differential Privacy [70.818129585404]
We experimentally demonstrate that networks, trained with differential privacy, in some settings might be even more vulnerable in comparison to non-private versions.
We study how the main ingredients of differentially private neural networks training, such as gradient clipping and noise addition, affect the robustness of the model.
arXiv Detail & Related papers (2020-12-14T18:59:24Z) - Differential Privacy of Hierarchical Census Data: An Optimization
Approach [53.29035917495491]
Census Bureaus are interested in releasing aggregate socio-economic data about a large population without revealing sensitive information about any individual.
Recent events have identified some of the privacy challenges faced by these organizations.
This paper presents a novel differential-privacy mechanism for releasing hierarchical counts of individuals.
arXiv Detail & Related papers (2020-06-28T18:19:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.