Modernizing Data Control: Making Personal Digital Data Mutually
Beneficial for Citizens and Industry
- URL: http://arxiv.org/abs/2012.08571v1
- Date: Tue, 15 Dec 2020 19:32:12 GMT
- Title: Modernizing Data Control: Making Personal Digital Data Mutually
Beneficial for Citizens and Industry
- Authors: Sujata Banerjee, Yiling Chen, Kobbi Nissim, David Parkes, Katie Siek,
and Lauren Wilcox
- Abstract summary: We are entering a new "data everywhere-anytime" era that pivots us from being tracked online to continuous tracking.
We create a lot of data, but who owns that data?
We look at major questions that policymakers should ask and things to consider when addressing data ownership concerns.
- Score: 13.479995363229877
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We are entering a new "data everywhere-anytime" era that pivots us from being
tracked online to continuous tracking as we move through our everyday lives. We
have smart devices in our homes, on our bodies, and around our communities that
collect data that is used to guide decisions that have a major impact on our
lives - from loans to job interviews and judicial rulings to health care
interventions. We create a lot of data, but who owns that data? How is it
shared? How will it be used? While the average person does not have a good
understanding of how the data is being used, they know that it carries risks
for them and society.
Although some people may believe they own their data, in reality, the problem
of understanding the myriad ways in which data is collected, shared, and used,
and the consequences of these uses is so complex that only a few people want to
manage their data themselves. Furthermore, much of the value in the data cannot
be extracted by individuals alone, as it lies in the connections and insights
garnered from (1) one's own personal data (is your fitness improving? Is your
home more energy efficient than the average home of this size?) and (2) one's
relationship with larger groups (demographic group voting blocks; friend
network influence on purchasing). But sometimes these insights have unintended
consequences for the person generating the data, especially in terms of loss of
privacy, unfairness, inappropriate inferences, information bias, manipulation,
and discrimination. There are also societal impacts, such as effects on speech
freedoms, political manipulation, and amplified harms to weakened and
underrepresented communities. To this end, we look at major questions that
policymakers should ask and things to consider when addressing these data
ownership concerns.
Related papers
- Differentially Private Data Release on Graphs: Inefficiencies and Unfairness [48.96399034594329]
This paper characterizes the impact of Differential Privacy on bias and unfairness in the context of releasing information about networks.
We consider a network release problem where the network structure is known to all, but the weights on edges must be released privately.
Our work provides theoretical foundations and empirical evidence into the bias and unfairness arising due to privacy in these networked decision problems.
arXiv Detail & Related papers (2024-08-08T08:37:37Z) - Insights from an experiment crowdsourcing data from thousands of US Amazon users: The importance of transparency, money, and data use [6.794366017852433]
This paper shares an innovative approach to crowdsourcing user data to collect otherwise inaccessible Amazon purchase histories, spanning 5 years, from more than 5000 US users.
We developed a data collection tool that prioritizes participant consent and includes an experimental study design.
Experiment results (N=6325) reveal both monetary incentives and transparency can significantly increase data sharing.
arXiv Detail & Related papers (2024-04-19T20:45:19Z) - Into the LAIONs Den: Investigating Hate in Multimodal Datasets [67.21783778038645]
This paper investigates the effect of scaling datasets on hateful content through a comparative audit of two datasets: LAION-400M and LAION-2B.
We found that hate content increased by nearly 12% with dataset scale, measured both qualitatively and quantitatively.
We also found that filtering dataset contents based on Not Safe For Work (NSFW) values calculated based on images alone does not exclude all the harmful content in alt-text.
arXiv Detail & Related papers (2023-11-06T19:00:05Z) - Data-Driven Dystopia: an uninterrupted breach of ethics [0.0]
The article presents instances of data breaches and data harvesting practices that violate user privacy.
It also explores the concept of "Weapons Of Math Destruction" (WMDs)
The article highlights the need for companies to take responsibility for safeguarding user information.
arXiv Detail & Related papers (2023-05-13T14:56:18Z) - Protecting User Privacy in Online Settings via Supervised Learning [69.38374877559423]
We design an intelligent approach to online privacy protection that leverages supervised learning.
By detecting and blocking data collection that might infringe on a user's privacy, we can restore a degree of digital privacy to the user.
arXiv Detail & Related papers (2023-04-06T05:20:16Z) - Contributing to Accessibility Datasets: Reflections on Sharing Study
Data by Blind People [14.625384963263327]
We present a pair of studies where 13 blind participants engage in data capturing activities.
We see how different factors influence blind participants' willingness to share study data as they assess risk-benefit tradeoffs.
The majority support sharing of their data to improve technology but also express concerns over commercial use, associated metadata, and the lack of transparency about the impact of their data.
arXiv Detail & Related papers (2023-03-09T00:42:18Z) - More Data Types More Problems: A Temporal Analysis of Complexity,
Stability, and Sensitivity in Privacy Policies [0.0]
Data brokers and data processors are part of a multi-billion-dollar industry that profits from collecting, buying, and selling consumer data.
Yet there is little transparency in the data collection industry which makes it difficult to understand what types of data are being collected, used, and sold.
arXiv Detail & Related papers (2023-02-17T15:21:24Z) - Second layer data governance for permissioned blockchains: the privacy
management challenge [58.720142291102135]
In pandemic situations, such as the COVID-19 and Ebola outbreak, the action related to sharing health data is crucial to avoid the massive infection and decrease the number of deaths.
In this sense, permissioned blockchain technology emerges to empower users to get their rights providing data ownership, transparency, and security through an immutable, unified, and distributed database ruled by smart contracts.
arXiv Detail & Related papers (2020-10-22T13:19:38Z) - Security and Privacy Preserving Deep Learning [2.322461721824713]
Massive data collection required for deep learning presents obvious privacy issues.
Users personal, highly sensitive data such as photos and voice recordings are kept indefinitely by the companies that collect it.
Deep neural networks are susceptible to various inference attacks as they remember information about their training data.
arXiv Detail & Related papers (2020-06-23T01:53:46Z) - A vision for global privacy bridges: Technical and legal measures for
international data markets [77.34726150561087]
Despite data protection laws and an acknowledged right to privacy, trading personal information has become a business equated with "trading oil"
An open conflict is arising between business demands for data and a desire for privacy.
We propose and test a vision of a personal information market with privacy.
arXiv Detail & Related papers (2020-05-13T13:55:50Z) - Contrastive Examples for Addressing the Tyranny of the Majority [83.93825214500131]
We propose to create a balanced training dataset, consisting of the original dataset plus new data points in which the group memberships are intervened.
We show that current generative adversarial networks are a powerful tool for learning these data points, called contrastive examples.
arXiv Detail & Related papers (2020-04-14T14:06:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.