Wisdom for the Crowd: Discoursive Power in Annotation Instructions for
Computer Vision
- URL: http://arxiv.org/abs/2105.10990v1
- Date: Sun, 23 May 2021 18:20:39 GMT
- Title: Wisdom for the Crowd: Discoursive Power in Annotation Instructions for
Computer Vision
- Authors: Milagros Miceli and Julian Posada
- Abstract summary: This paper focuses on the experiences of Argentinian and Venezuelan data annotators.
Our findings indicate that annotation instructions reflect worldviews imposed on workers and, through their labor, on datasets.
This configuration presents a form of commodified labor that perpetuates power asymmetries while reinforcing social inequalities.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Developers of computer vision algorithms outsource some of the labor involved
in annotating training data through business process outsourcing companies and
crowdsourcing platforms. Many data annotators are situated in the Global South
and are considered independent contractors. This paper focuses on the
experiences of Argentinian and Venezuelan annotation workers. Through
qualitative methods, we explore the discourses encoded in the task instructions
that these workers follow to annotate computer vision datasets. Our preliminary
findings indicate that annotation instructions reflect worldviews imposed on
workers and, through their labor, on datasets. Moreover, we observe that
for-profit goals drive task instructions and that managers and algorithms make
sure annotations are done according to requesters' commands. This configuration
presents a form of commodified labor that perpetuates power asymmetries while
reinforcing social inequalities and is compelled to reproduce them into
datasets and, subsequently, in computer vision systems.
Related papers
- Global Inequalities in the Production of Artificial Intelligence: A Four-Country Study on Data Work [0.0]
Labor plays a major, albeit largely unrecognized role in the development of artificial intelligence.
Online platforms and networks of subcontractors recruit data workers to execute tasks in the shadow of AI production.
This study unveils the resulting complexities by comparing the working conditions and the profiles of data workers in Venezuela, Brazil, Madagascar, and as an example of a richer country, France.
arXiv Detail & Related papers (2024-10-18T07:23:17Z) - From Pixels to Insights: A Survey on Automatic Chart Understanding in the Era of Large Foundation Models [98.41645229835493]
Data visualization in the form of charts plays a pivotal role in data analysis, offering critical insights and aiding in informed decision-making.
Large foundation models, such as large language models, have revolutionized various natural language processing tasks.
This survey paper serves as a comprehensive resource for researchers and practitioners in the fields of natural language processing, computer vision, and data analysis.
arXiv Detail & Related papers (2024-03-18T17:57:09Z) - Vision-Language Models Provide Promptable Representations for Reinforcement Learning [67.40524195671479]
We propose a novel approach that uses the vast amounts of general and indexable world knowledge encoded in vision-language models (VLMs) pre-trained on Internet-scale data for embodied reinforcement learning (RL)
We show that our approach can use chain-of-thought prompting to produce representations of common-sense semantic reasoning, improving policy performance in novel scenes by 1.5 times.
arXiv Detail & Related papers (2024-02-05T00:48:56Z) - InstructCV: Instruction-Tuned Text-to-Image Diffusion Models as Vision Generalists [66.85125112199898]
We develop a unified language interface for computer vision tasks that abstracts away task-specific design choices.
Our model, dubbed InstructCV, performs competitively compared to other generalist and task-specific vision models.
arXiv Detail & Related papers (2023-09-30T14:26:43Z) - What and How of Machine Learning Transparency: Building Bespoke
Explainability Tools with Interoperable Algorithmic Components [77.87794937143511]
This paper introduces a collection of hands-on training materials for explaining data-driven predictive models.
These resources cover the three core building blocks of this technique: interpretable representation composition, data sampling and explanation generation.
arXiv Detail & Related papers (2022-09-08T13:33:25Z) - The Data-Production Dispositif [0.0]
This paper investigates outsourced machine learning data work in Latin America by studying three platforms in Venezuela and a BPO in Argentina.
We lean on the Foucauldian notion of dispositif to define the data-production dispositif as an ensemble of discourses, actions, and objects strategically disposed to (re)produce power/knowledge relations in data and labor.
We conclude by stressing the importance of counteracting the data-production dispositif by fighting alienation and precarization, and empowering data workers to become assets in the quest for high-quality data.
arXiv Detail & Related papers (2022-05-24T10:51:05Z) - Re-Examining Human Annotations for Interpretable NLP [80.81532239566992]
We conduct controlled experiments using crowd-sourced websites on two widely used datasets in Interpretable NLP.
We compare the annotation results obtained from recruiting workers satisfying different levels of qualification.
Our results reveal that the annotation quality is highly subject to the workers' qualification, and workers can be guided to provide certain annotations by the instructions.
arXiv Detail & Related papers (2022-04-10T02:27:30Z) - Do Datasets Have Politics? Disciplinary Values in Computer Vision
Dataset Development [6.182409582844314]
We collect a corpus of about 500 computer vision datasets, from which we sampled 114 dataset publications across different vision tasks.
We discuss how computer vision datasets authors value efficiency at the expense of care; universality at the expense of contextuality; and model work at the expense of data work.
We conclude with suggestions on how to better incorporate silenced values into the dataset creation and curation process.
arXiv Detail & Related papers (2021-08-09T19:07:58Z) - Reinforcement Learning, Bit by Bit [27.66567077899924]
Reinforcement learning agents have demonstrated remarkable achievements in simulated environments.
Data efficiency poses an impediment to carrying this success over to real environments.
We discuss concepts and regret analysis that together offer principled guidance.
arXiv Detail & Related papers (2021-03-06T06:37:46Z) - Reasoning over Vision and Language: Exploring the Benefits of
Supplemental Knowledge [59.87823082513752]
This paper investigates the injection of knowledge from general-purpose knowledge bases (KBs) into vision-and-language transformers.
We empirically study the relevance of various KBs to multiple tasks and benchmarks.
The technique is model-agnostic and can expand the applicability of any vision-and-language transformer with minimal computational overhead.
arXiv Detail & Related papers (2021-01-15T08:37:55Z) - Bringing the People Back In: Contesting Benchmark Machine Learning
Datasets [11.00769651520502]
We outline a research program - a genealogy of machine learning data - for investigating how and why these datasets have been created.
We describe the ways in which benchmark datasets in machine learning operate as infrastructure and pose four research questions for these datasets.
arXiv Detail & Related papers (2020-07-14T23:22:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.