Social and Governance Implications of Improved Data Efficiency
- URL: http://arxiv.org/abs/2001.05068v1
- Date: Tue, 14 Jan 2020 22:26:12 GMT
- Title: Social and Governance Implications of Improved Data Efficiency
- Authors: Aaron D. Tucker, Markus Anderljung, and Allan Dafoe
- Abstract summary: This paper explores the social-economic impact of increased data efficiency.
We find that the effect on privacy, data markets, robustness, and misuse are complex.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Many researchers work on improving the data efficiency of machine learning.
What would happen if they succeed? This paper explores the social-economic
impact of increased data efficiency. Specifically, we examine the intuition
that data efficiency will erode the barriers to entry protecting incumbent
data-rich AI firms, exposing them to more competition from data-poor firms. We
find that this intuition is only partially correct: data efficiency makes it
easier to create ML applications, but large AI firms may have more to gain from
higher performing AI systems. Further, we find that the effect on privacy, data
markets, robustness, and misuse are complex. For example, while it seems
intuitive that misuse risk would increase along with data efficiency -- as more
actors gain access to any level of capability -- the net effect crucially
depends on how much defensive measures are improved. More investigation into
data efficiency, as well as research into the "AI production function", will be
key to understanding the development of the AI industry and its societal
impacts.
Related papers
- AI Cap-and-Trade: Efficiency Incentives for Accessibility and Sustainability [16.11189838235793]
We argue for research into, and implementation of, market-based methods that incentivize AI efficiency.<n>As a call to action, we propose a cap-and-trade system for AI.
arXiv Detail & Related papers (2026-01-27T18:53:21Z) - When Assurance Undermines Intelligence: The Efficiency Costs of Data Governance in AI-Enabled Labor Markets [5.3700224653806865]
We show that restricting data use significantly reduced GenAI efficiency, leading to lower matching rates, higher employee turnover, and heightened labor market frictions.<n>Our findings reveal the unintended efficiency costs of well-intentioned data governance and highlight that information assurance, while essential for trust, can undermine intelligence-driven efficiency when misaligned with AI system design.
arXiv Detail & Related papers (2025-11-02T05:35:37Z) - The SPACE of AI: Real-World Lessons on AI's Impact on Developers [0.807084206814932]
We study how developers perceive AI's influence across the dimensions of the SPACE framework: Satisfaction, Performance, Activity, Collaboration and Efficiency.<n>We find that AI is broadly adopted and widely seen as enhancing productivity, particularly for routine tasks.<n>Developers report increased efficiency and satisfaction, with less evidence of impact on collaboration.
arXiv Detail & Related papers (2025-07-31T21:45:54Z) - AI, Humans, and Data Science: Optimizing Roles Across Workflows and the Workforce [0.0]
We consider the potential and limitation of analytic, generative, and agentic AI to augment data scientists or take on tasks traditionally done by human analysts and researchers.<n>Just as earlier eras of survey analysis created issues when the increased ease of using statistical software allowed researchers to conduct analyses they did not fully understand, the new AI tools may create similar but larger risks.
arXiv Detail & Related papers (2025-07-15T17:59:06Z) - Addressing Bias in LLMs: Strategies and Application to Fair AI-based Recruitment [49.81946749379338]
This work seeks to analyze the capacity of Transformers-based systems to learn demographic biases present in the data.<n>We propose a privacy-enhancing framework to reduce gender information from the learning pipeline as a way to mitigate biased behaviors in the final tools.
arXiv Detail & Related papers (2025-06-13T15:29:43Z) - Raising the Stakes: Performance Pressure Improves AI-Assisted Decision Making [57.53469908423318]
We show the effects of performance pressure on AI advice reliance when laypeople complete a common AI-assisted task.
We find that when the stakes are high, people use AI advice more appropriately than when stakes are lower, regardless of the presence of an AI explanation.
arXiv Detail & Related papers (2024-10-21T22:39:52Z) - Ethical AI in Retail: Consumer Privacy and Fairness [0.0]
The adoption of artificial intelligence (AI) in retail has significantly transformed the industry, enabling more personalized services and efficient operations.
However, the rapid implementation of AI technologies raises ethical concerns, particularly regarding consumer privacy and fairness.
This study aims to analyze the ethical challenges of AI applications in retail, explore ways retailers can implement AI technologies ethically while remaining competitive, and provide recommendations on ethical AI practices.
arXiv Detail & Related papers (2024-10-20T12:00:14Z) - Hype, Sustainability, and the Price of the Bigger-is-Better Paradigm in AI [67.58673784790375]
We argue that the 'bigger is better' AI paradigm is not only fragile scientifically, but comes with undesirable consequences.
First, it is not sustainable, as, despite efficiency improvements, its compute demands increase faster than model performance.
Second, it implies focusing on certain problems at the expense of others, leaving aside important applications, e.g. health, education, or the climate.
arXiv Detail & Related papers (2024-09-21T14:43:54Z) - Generative AI for Secure and Privacy-Preserving Mobile Crowdsensing [74.58071278710896]
generative AI has attracted much attention from both academic and industrial fields.
Secure and privacy-preserving mobile crowdsensing (SPPMCS) has been widely applied in data collection/ acquirement.
arXiv Detail & Related papers (2024-05-17T04:00:58Z) - The Frontier of Data Erasure: Machine Unlearning for Large Language Models [56.26002631481726]
Large Language Models (LLMs) are foundational to AI advancements.
LLMs pose risks by potentially memorizing and disseminating sensitive, biased, or copyrighted information.
Machine unlearning emerges as a cutting-edge solution to mitigate these concerns.
arXiv Detail & Related papers (2024-03-23T09:26:15Z) - On Responsible Machine Learning Datasets with Fairness, Privacy, and Regulatory Norms [56.119374302685934]
There have been severe concerns over the trustworthiness of AI technologies.
Machine and deep learning algorithms depend heavily on the data used during their development.
We propose a framework to evaluate the datasets through a responsible rubric.
arXiv Detail & Related papers (2023-10-24T14:01:53Z) - The Dimensions of Data Labor: A Road Map for Researchers, Activists, and
Policymakers to Empower Data Producers [14.392208044851976]
Data producers have little say in what data is captured, how it is used, or who it benefits.
Organizations with the ability to access and process this data, e.g. OpenAI and Google, possess immense power in shaping the technology landscape.
By synthesizing related literature that reconceptualizes the production of data for computing as data labor'', we outline opportunities for researchers, policymakers, and activists to empower data producers.
arXiv Detail & Related papers (2023-05-22T17:11:22Z) - Towards Generalizable Data Protection With Transferable Unlearnable
Examples [50.628011208660645]
We present a novel, generalizable data protection method by generating transferable unlearnable examples.
To the best of our knowledge, this is the first solution that examines data privacy from the perspective of data distribution.
arXiv Detail & Related papers (2023-05-18T04:17:01Z) - Data-Centric Green AI: An Exploratory Empirical Study [6.4265933507484]
We investigate the impact of data-centric approaches on AI energy efficiency.
Our results show evidence that, by exclusively conducting modifications on datasets, energy consumption can be drastically reduced.
Our results call for a research agenda that focuses on data-centric techniques to further enable and democratize Green AI.
arXiv Detail & Related papers (2022-04-06T12:22:43Z) - AI Assurance using Causal Inference: Application to Public Policy [0.0]
Most AI approaches can only be represented as "black boxes" and suffer from the lack of transparency.
It is crucial not only to develop effective and robust AI systems, but to make sure their internal processes are explainable and fair.
arXiv Detail & Related papers (2021-12-01T16:03:06Z) - Provably Efficient Causal Reinforcement Learning with Confounded
Observational Data [135.64775986546505]
We study how to incorporate the dataset (observational data) collected offline, which is often abundantly available in practice, to improve the sample efficiency in the online setting.
We propose the deconfounded optimistic value iteration (DOVI) algorithm, which incorporates the confounded observational data in a provably efficient manner.
arXiv Detail & Related papers (2020-06-22T14:49:33Z) - Quantifying the Performance of Federated Transfer Learning [7.1423970352437385]
Federated Transfer Learning (FTL) is a solution to share data without violating data privacy.
FTL uses transfer learning techniques to utilize data from different sources for training.
Our paper tries to answer this question by quantitatively measuring a real-world FTL implementation on Google Cloud.
arXiv Detail & Related papers (2019-12-30T03:10:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.