Exploring Data Pipelines through the Process Lens: a Reference Model
forComputer Vision
- URL: http://arxiv.org/abs/2107.01824v1
- Date: Mon, 5 Jul 2021 07:15:57 GMT
- Title: Exploring Data Pipelines through the Process Lens: a Reference Model
forComputer Vision
- Authors: Agathe Balayn, Bogdan Kulynych, Seda Guerses
- Abstract summary: We argue that we could further systematize our analysis of harms by examining CV data pipelines through a process-oriented lens.
As a step towards cultivating a process-oriented lens, we embarked on an empirical study of CV data pipelines.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Researchers have identified datasets used for training computer vision (CV)
models as an important source of hazardous outcomes, and continue to examine
popular CV datasets to expose their harms. These works tend to treat datasets
as objects, or focus on particular steps in data production pipelines. We argue
here that we could further systematize our analysis of harms by examining CV
data pipelines through a process-oriented lens that captures the creation, the
evolution and use of these datasets. As a step towards cultivating a
process-oriented lens, we embarked on an empirical study of CV data pipelines
informed by the field of method engineering. We present here a preliminary
result: a reference model of CV data pipelines. Besides exploring the questions
that this endeavor raises, we discuss how the process lens could support
researchers in discovering understudied issues, and could help practitioners in
making their processes more transparent.
Related papers
- Enabling Advanced Land Cover Analytics: An Integrated Data Extraction Pipeline for Predictive Modeling with the Dynamic World Dataset [1.3757956340051605]
We present a flexible and efficient end to end pipeline for working with the Dynamic World dataset.
This includes a pre-processing and representation framework which tackles noise removal, efficient extraction of large amounts of data, and re-representation of LULC data.
To demonstrate the power of our pipeline, we use it to extract data for an urbanization prediction problem and build a suite of machine learning models with excellent performance.
arXiv Detail & Related papers (2024-10-11T16:13:01Z) - A Systematic Review of Available Datasets in Additive Manufacturing [56.684125592242445]
In-situ monitoring incorporating visual and other sensor technologies allows the collection of extensive datasets during the Additive Manufacturing process.
These datasets have potential for determining the quality of the manufactured output and the detection of defects through the use of Machine Learning.
This systematic review investigates the availability of open image-based datasets originating from AM processes that align with a number of pre-defined selection criteria.
arXiv Detail & Related papers (2024-01-27T16:13:32Z) - ALP: Action-Aware Embodied Learning for Perception [60.64801970249279]
We introduce Action-Aware Embodied Learning for Perception (ALP)
ALP incorporates action information into representation learning through a combination of optimizing a reinforcement learning policy and an inverse dynamics prediction objective.
We show that ALP outperforms existing baselines in several downstream perception tasks.
arXiv Detail & Related papers (2023-06-16T21:51:04Z) - Demonstration of InsightPilot: An LLM-Empowered Automated Data
Exploration System [48.62158108517576]
We introduce InsightPilot, an automated data exploration system designed to simplify the data exploration process.
InsightPilot automatically selects appropriate analysis intents, such as understanding, summarizing, and explaining.
In brief, an IQuery is an abstraction and automation of data analysis operations, which mimics the approach of data analysts.
arXiv Detail & Related papers (2023-04-02T07:27:49Z) - TRoVE: Transforming Road Scene Datasets into Photorealistic Virtual
Environments [84.6017003787244]
This work proposes a synthetic data generation pipeline to address the difficulties and domain-gaps present in simulated datasets.
We show that using annotations and visual cues from existing datasets, we can facilitate automated multi-modal data generation.
arXiv Detail & Related papers (2022-08-16T20:46:08Z) - A Tour of Visualization Techniques for Computer Vision Datasets [4.5916483318867]
We survey a number of data visualization techniques for analyzing Computer Vision (CV) datasets.
These techniques help us understand properties and latent patterns in such data, by applying dataset-level analysis.
We present various examples of how such analysis helps predict the potential impact of the dataset properties on CV models.
arXiv Detail & Related papers (2022-04-19T01:04:28Z) - Omnidata: A Scalable Pipeline for Making Multi-Task Mid-Level Vision
Datasets from 3D Scans [103.92680099373567]
This paper introduces a pipeline to parametrically sample and render multi-task vision datasets from comprehensive 3D scans from the real world.
Changing the sampling parameters allows one to "steer" the generated datasets to emphasize specific information.
Common architectures trained on a generated starter dataset reached state-of-the-art performance on multiple common vision tasks and benchmarks.
arXiv Detail & Related papers (2021-10-11T04:21:46Z) - On The State of Data In Computer Vision: Human Annotations Remain
Indispensable for Developing Deep Learning Models [0.0]
High-quality labeled datasets play a crucial role in fueling the development of machine learning (ML)
Since the emergence of the ImageNet dataset and the AlexNet model in 2012, the size of new open-source labeled vision datasets has remained roughly constant.
Only a minority of publications in the computer vision community tackle supervised learning on datasets that are orders of magnitude larger than Imagenet.
arXiv Detail & Related papers (2021-07-31T00:08:21Z) - From ImageNet to Image Classification: Contextualizing Progress on
Benchmarks [99.19183528305598]
We study how specific design choices in the ImageNet creation process impact the fidelity of the resulting dataset.
Our analysis pinpoints how a noisy data collection pipeline can lead to a systematic misalignment between the resulting benchmark and the real-world task it serves as a proxy for.
arXiv Detail & Related papers (2020-05-22T17:39:16Z) - A Common Operating Picture Framework Leveraging Data Fusion and Deep
Learning [0.7348448478819135]
We present a data fusion framework for accelerating solutions for Processing, Exploitation, and Dissemination.
Our platform is a collection of services that extract information from several data sources by leveraging deep learning and other means of processing.
In our first iteration we have focused on visual data (FMV, WAMI, CCTV/PTZ-Cameras, open source video, etc.) and AIS data streams (satellite and terrestrial sources)
arXiv Detail & Related papers (2020-01-16T18:32:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.