When Assurance Undermines Intelligence: The Efficiency Costs of Data Governance in AI-Enabled Labor Markets
- URL: http://arxiv.org/abs/2511.01923v1
- Date: Sun, 02 Nov 2025 05:35:37 GMT
- Title: When Assurance Undermines Intelligence: The Efficiency Costs of Data Governance in AI-Enabled Labor Markets
- Authors: Lei Chen, Chaoyue Gao, Alvin Leung, Xiaoning Wang,
- Abstract summary: We show that restricting data use significantly reduced GenAI efficiency, leading to lower matching rates, higher employee turnover, and heightened labor market frictions.<n>Our findings reveal the unintended efficiency costs of well-intentioned data governance and highlight that information assurance, while essential for trust, can undermine intelligence-driven efficiency when misaligned with AI system design.
- Score: 5.3700224653806865
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Generative artificial intelligence (GenAI) like Large Language Model (LLM) is increasingly integrated into digital platforms to enhance information access, deliver personalized experiences, and improve matching efficiency. However, these algorithmic advancements rely heavily on large-scale user data, creating a fundamental tension between information assurance-the protection, integrity, and responsible use of privacy data-and artificial intelligence-the learning capacity and predictive accuracy of models. We examine this assurance-intelligence trade-off in the context of LinkedIn, leveraging a regulatory intervention that suspended the use of user data for model training in Hong Kong. Using large-scale employment and job posting data from Revelio Labs and a Difference-in-Differences design, we show that restricting data use significantly reduced GenAI efficiency, leading to lower matching rates, higher employee turnover, and heightened labor market frictions. These effects were especially pronounced for small and fast-growing firms that rely heavily on AI for talent acquisition. Our findings reveal the unintended efficiency costs of well-intentioned data governance and highlight that information assurance, while essential for trust, can undermine intelligence-driven efficiency when misaligned with AI system design. This study contributes to emerging research on AI governance and digital platform by theorizing data assurance as an institutional complement-and potential constraint-to GenAI efficacy in data-intensive environments.
Related papers
- What's the next frontier for Data-centric AI? Data Savvy Agents [71.76058707995398]
We argue that data-savvy capabilities should be a top priority in the design of agentic systems.<n>We propose four key capabilities to realize this vision: Proactive data acquisition, Sophisticated data processing, Interactive test data synthesis, and Continual adaptation.
arXiv Detail & Related papers (2025-11-02T17:09:29Z) - Improving AI Efficiency in Data Centres by Power Dynamic Response [74.12165648170894]
The steady growth of artificial intelligence (AI) has accelerated in the recent years, facilitated by the development of sophisticated models.<n> Ensuring robust and reliable power infrastructures is fundamental to take advantage of the full potential of AI.<n>However, AI data centres are extremely hungry for power, putting the problem of their power management in the spotlight.
arXiv Detail & Related papers (2025-10-13T08:08:21Z) - Autonomous Data Agents: A New Opportunity for Smart Data [50.02229219403014]
Report argues that DataAgents represent a paradigm shift toward autonomous data-to-knowledge systems.<n>DataAgents transform complex and unstructured data into coherent and actionable knowledge.<n>We first examine why the convergence of agentic AI and data-to-knowledge systems has emerged as a critical trend.
arXiv Detail & Related papers (2025-09-23T06:46:41Z) - From Machine Learning to Machine Unlearning: Complying with GDPR's Right to be Forgotten while Maintaining Business Value of Predictive Models [9.380866972744633]
This work develops a holistic machine learning-to-unlearning framework, called Ensemble-based iTerative Information Distillation (ETID)<n>ETID incorporates a new ensemble learning method to build an accurate predictive model that can facilitate handling data erasure requests.<n>We also introduce an innovative distillation-based unlearning method tailored to the constructed ensemble model to enable efficient and effective data erasure.
arXiv Detail & Related papers (2024-11-26T05:42:46Z) - Accuracy of training data and model outputs in Generative AI: CREATe Response to the Information Commissioner Office Consultation [6.699484354380871]
CREATe welcomes the ICO call for evidence on the accuracy of Generative AI.
We are happy to highlight aspects of data protection law and AI regulation that we believe should receive attention.
arXiv Detail & Related papers (2024-05-30T10:34:49Z) - On Responsible Machine Learning Datasets with Fairness, Privacy, and Regulatory Norms [56.119374302685934]
There have been severe concerns over the trustworthiness of AI technologies.
Machine and deep learning algorithms depend heavily on the data used during their development.
We propose a framework to evaluate the datasets through a responsible rubric.
arXiv Detail & Related papers (2023-10-24T14:01:53Z) - Leveraging Centric Data Federated Learning Using Blockchain For
Integrity Assurance [14.347917009290814]
We propose a data-centric federated learning architecture leveraged by a public blockchain and smart contracts.
Our proposed solution provides a virtual public marketplace where developers, data scientists, and AI-engineer can publish their models.
We enhance data quality and integrity through an incentive mechanism that rewards contributors for data contribution and verification.
arXiv Detail & Related papers (2022-06-09T19:06:05Z) - Adaptive cognitive fit: Artificial intelligence augmented management of
information facets and representations [62.997667081978825]
Explosive growth in big data technologies and artificial intelligence [AI] applications have led to increasing pervasiveness of information facets.
Information facets, such as equivocality and veracity, can dominate and significantly influence human perceptions of information.
We suggest that artificially intelligent technologies that can adapt information representations to overcome cognitive limitations are necessary.
arXiv Detail & Related papers (2022-04-25T02:47:25Z) - AI Assurance using Causal Inference: Application to Public Policy [0.0]
Most AI approaches can only be represented as "black boxes" and suffer from the lack of transparency.
It is crucial not only to develop effective and robust AI systems, but to make sure their internal processes are explainable and fair.
arXiv Detail & Related papers (2021-12-01T16:03:06Z) - Trustworthy AI [75.99046162669997]
Brittleness to minor adversarial changes in the input data, ability to explain the decisions, address the bias in their training data, are some of the most prominent limitations.
We propose the tutorial on Trustworthy AI to address six critical issues in enhancing user and public trust in AI systems.
arXiv Detail & Related papers (2020-11-02T20:04:18Z) - Social and Governance Implications of Improved Data Efficiency [0.0]
This paper explores the social-economic impact of increased data efficiency.
We find that the effect on privacy, data markets, robustness, and misuse are complex.
arXiv Detail & Related papers (2020-01-14T22:26:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.