The Values Encoded in Machine Learning Research
- URL: http://arxiv.org/abs/2106.15590v1
- Date: Tue, 29 Jun 2021 17:24:14 GMT
- Title: The Values Encoded in Machine Learning Research
- Authors: Abeba Birhane, Pratyusha Kalluri, Dallas Card, William Agnew, Ravit
Dotan, Michelle Bao
- Abstract summary: We analyze 100 highly cited machine learning papers published at premier conferences, ICML and NeurIPS.
We identify 67 values that are uplifted in machine learning research.
We find increasingly close ties between these highly cited papers and tech companies and elite universities.
- Score: 6.11644847221881
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Machine learning (ML) currently exerts an outsized influence on the world,
increasingly affecting communities and institutional practices. It is therefore
critical that we question vague conceptions of the field as value-neutral or
universally beneficial, and investigate what specific values the field is
advancing. In this paper, we present a rigorous examination of the values of
the field by quantitatively and qualitatively analyzing 100 highly cited ML
papers published at premier ML conferences, ICML and NeurIPS. We annotate key
features of papers which reveal their values: how they justify their choice of
project, which aspects they uplift, their consideration of potential negative
consequences, and their institutional affiliations and funding sources. We find
that societal needs are typically very loosely connected to the choice of
project, if mentioned at all, and that consideration of negative consequences
is extremely rare. We identify 67 values that are uplifted in machine learning
research, and, of these, we find that papers most frequently justify and assess
themselves based on performance, generalization, efficiency, researcher
understanding, novelty, and building on previous work. We present extensive
textual evidence and analysis of how these values are operationalized. Notably,
we find that each of these top values is currently being defined and applied
with assumptions and implications generally supporting the centralization of
power. Finally, we find increasingly close ties between these highly cited
papers and tech companies and elite universities.
Related papers
- RelevAI-Reviewer: A Benchmark on AI Reviewers for Survey Paper Relevance [0.8089605035945486]
We propose RelevAI-Reviewer, an automatic system that conceptualizes the task of survey paper review as a classification problem.
We introduce a novel dataset comprised of 25,164 instances. Each instance contains one prompt and four candidate papers, each varying in relevance to the prompt.
We develop a machine learning (ML) model capable of determining the relevance of each paper and identifying the most pertinent one.
arXiv Detail & Related papers (2024-06-13T06:42:32Z) - Responsible AI Considerations in Text Summarization Research: A Review
of Current Practices [89.85174013619883]
We focus on text summarization, a common NLP task largely overlooked by the responsible AI community.
We conduct a multi-round qualitative analysis of 333 summarization papers from the ACL Anthology published between 2020-2022.
We focus on how, which, and when responsible AI issues are covered, which relevant stakeholders are considered, and mismatches between stated and realized research goals.
arXiv Detail & Related papers (2023-11-18T15:35:36Z) - Investigating Fairness Disparities in Peer Review: A Language Model
Enhanced Approach [77.61131357420201]
We conduct a thorough and rigorous study on fairness disparities in peer review with the help of large language models (LMs)
We collect, assemble, and maintain a comprehensive relational database for the International Conference on Learning Representations (ICLR) conference from 2017 to date.
We postulate and study fairness disparities on multiple protective attributes of interest, including author gender, geography, author, and institutional prestige.
arXiv Detail & Related papers (2022-11-07T16:19:42Z) - Fairness in Recommender Systems: Research Landscape and Future
Directions [119.67643184567623]
We review the concepts and notions of fairness that were put forward in the area in the recent past.
We present an overview of how research in this field is currently operationalized.
Overall, our analysis of recent works points to certain research gaps.
arXiv Detail & Related papers (2022-05-23T08:34:25Z) - The Forgotten Margins of AI Ethics [4.150076021042422]
We read and annotated peer-reviewed papers published over the past four years in two premier conferences: FAccT and AIES.
We note that although the goals of the majority of FAccT and AIES papers were often commendable, their consideration of the negative impacts of AI on traditionally marginalized groups remained shallow.
arXiv Detail & Related papers (2022-05-09T12:16:29Z) - Metaethical Perspectives on 'Benchmarking' AI Ethics [81.65697003067841]
Benchmarks are seen as the cornerstone for measuring technical progress in Artificial Intelligence (AI) research.
An increasingly prominent research area in AI is ethics, which currently has no set of benchmarks nor commonly accepted way for measuring the 'ethicality' of an AI system.
We argue that it makes more sense to talk about 'values' rather than 'ethics' when considering the possible actions of present and future AI systems.
arXiv Detail & Related papers (2022-04-11T14:36:39Z) - Algorithmic Fairness Datasets: the Story so Far [68.45921483094705]
Data-driven algorithms are studied in diverse domains to support critical decisions, directly impacting people's well-being.
A growing community of researchers has been investigating the equity of existing algorithms and proposing novel ones, advancing the understanding of risks and opportunities of automated decision-making for historically disadvantaged populations.
Progress in fair Machine Learning hinges on data, which can be appropriately used only if adequately documented.
Unfortunately, the algorithmic fairness community suffers from a collective data documentation debt caused by a lack of information on specific resources (opacity) and scatteredness of available information (sparsity)
arXiv Detail & Related papers (2022-02-03T17:25:46Z) - Affect Analysis in-the-wild: Valence-Arousal, Expressions, Action Units
and a Unified Framework [83.21732533130846]
The paper focuses on large in-the-wild databases, i.e., Aff-Wild and Aff-Wild2.
It presents the design of two classes of deep neural networks trained with these databases.
A novel multi-task and holistic framework is presented which is able to jointly learn and effectively generalize and perform affect recognition.
arXiv Detail & Related papers (2021-03-29T17:36:20Z) - Ethical Considerations and Statistical Analysis of Industry Involvement
in Machine Learning Research [3.4773470589069473]
We have examined all papers of the main ML conferences NeurIPS, CVPR, and ICML of the last 5 years.
Our statistical approach focuses on conflicts of interest, innovation and gender equality.
arXiv Detail & Related papers (2020-06-08T12:51:38Z) - Steps Towards Value-Aligned Systems [0.0]
Algorithmic (including AI/ML) decision-making artifacts are an established and growing part of our decision-making ecosystem.
Current literature is full of examples of how individual artifacts violate societal norms and expectations.
This discussion argues for a more structured systems-level approach for assessing value-alignment in sociotechnical systems.
arXiv Detail & Related papers (2020-02-10T22:47:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.