Comprehensive and Efficient Data Labeling via Adaptive Model Scheduling
- URL: http://arxiv.org/abs/2002.05520v1
- Date: Sat, 8 Feb 2020 03:54:39 GMT
- Title: Comprehensive and Efficient Data Labeling via Adaptive Model Scheduling
- Authors: Mu Yuan, Lan Zhang, Xiang-Yang Li, Hui Xiong
- Abstract summary: In certain applications, such as image retrieval platforms and photo album management apps, it is often required to execute a collection of models to obtain sufficient labels.
We propose an Adaptive Model Scheduling framework, consisting of 1) a deep reinforcement learning-based approach to predict the value of untrivial models by mining semantic relationship among diverse models, and 2) two algorithms to adaptively schedule the model execution order under a deadline or deadline-memory constraints respectively.
Our design could save around 53% execution time without loss of any valuable labels.
- Score: 25.525371500391568
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Labeling data (e.g., labeling the people, objects, actions and scene in
images) comprehensively and efficiently is a widely needed but challenging
task. Numerous models were proposed to label various data and many approaches
were designed to enhance the ability of deep learning models or accelerate
them. Unfortunately, a single machine-learning model is not powerful enough to
extract various semantic information from data. Given certain applications,
such as image retrieval platforms and photo album management apps, it is often
required to execute a collection of models to obtain sufficient labels. With
limited computing resources and stringent delay, given a data stream and a
collection of applicable resource-hungry deep-learning models, we design a
novel approach to adaptively schedule a subset of these models to execute on
each data item, aiming to maximize the value of the model output (e.g., the
number of high-confidence labels). Achieving this lofty goal is nontrivial
since a model's output on any data item is content-dependent and unknown until
we execute it. To tackle this, we propose an Adaptive Model Scheduling
framework, consisting of 1) a deep reinforcement learning-based approach to
predict the value of unexecuted models by mining semantic relationship among
diverse models, and 2) two heuristic algorithms to adaptively schedule the
model execution order under a deadline or deadline-memory constraints
respectively. The proposed framework doesn't require any prior knowledge of the
data, which works as a powerful complement to existing model optimization
technologies. We conduct extensive evaluations on five diverse image datasets
and 30 popular image labeling models to demonstrate the effectiveness of our
design: our design could save around 53\% execution time without loss of any
valuable labels.
Related papers
- Iterative Object Count Optimization for Text-to-image Diffusion Models [59.03672816121209]
Current models, which learn from image-text pairs, inherently struggle with counting.
We propose optimizing the generated image based on a counting loss derived from a counting model that aggregates an object's potential.
We evaluate the generation of various objects and show significant improvements in accuracy.
arXiv Detail & Related papers (2024-08-21T15:51:46Z) - Model Selection with Model Zoo via Graph Learning [45.30615308692713]
We introduce TransferGraph, a novel framework that reformulates model selection as a graph learning problem.
We demonstrate TransferGraph's effectiveness in capturing essential model-dataset relationships, yielding up to a 32% improvement in correlation between predicted performance and the actual fine-tuning results compared to the state-of-the-art methods.
arXiv Detail & Related papers (2024-04-05T09:50:00Z) - A Two-Phase Recall-and-Select Framework for Fast Model Selection [13.385915962994806]
We propose a two-phase (coarse-recall and fine-selection) model selection framework.
It aims to enhance the efficiency of selecting a robust model by leveraging the models' training performances on benchmark datasets.
It has been demonstrated that the proposed methodology facilitates the selection of a high-performing model at a rate about 3x times faster than conventional baseline methods.
arXiv Detail & Related papers (2024-03-28T14:44:44Z) - Modeling Collaborator: Enabling Subjective Vision Classification With Minimal Human Effort via LLM Tool-Use [14.2527771630478]
We propose a new framework that alleviates manual effort by replacing human labeling with natural language interactions.
Our framework eliminates the need for crowd-sourced annotations.
Our trained models outperform traditional Agile Modeling as well as state-of-the-art zero-shot classification models.
arXiv Detail & Related papers (2024-03-05T03:34:11Z) - Data-efficient Large Vision Models through Sequential Autoregression [58.26179273091461]
We develop an efficient, autoregression-based vision model on a limited dataset.
We demonstrate how this model achieves proficiency in a spectrum of visual tasks spanning both high-level and low-level semantic understanding.
Our empirical evaluations underscore the model's agility in adapting to various tasks, heralding a significant reduction in the parameter footprint.
arXiv Detail & Related papers (2024-02-07T13:41:53Z) - A Simple and Efficient Baseline for Data Attribution on Images [107.12337511216228]
Current state-of-the-art approaches require a large ensemble of as many as 300,000 models to accurately attribute model predictions.
In this work, we focus on a minimalist baseline, utilizing the feature space of a backbone pretrained via self-supervised learning to perform data attribution.
Our method is model-agnostic and scales easily to large datasets.
arXiv Detail & Related papers (2023-11-03T17:29:46Z) - Revealing the Underlying Patterns: Investigating Dataset Similarity,
Performance, and Generalization [0.0]
Supervised deep learning models require significant amount of labeled data to achieve an acceptable performance on a specific task.
We establish image-image, dataset-dataset, and image-dataset distances to gain insights into the model's behavior.
arXiv Detail & Related papers (2023-08-07T13:35:53Z) - An Efficient General-Purpose Modular Vision Model via Multi-Task
Heterogeneous Training [79.78201886156513]
We present a model that can perform multiple vision tasks and can be adapted to other downstream tasks efficiently.
Our approach achieves comparable results to single-task state-of-the-art models and demonstrates strong generalization on downstream tasks.
arXiv Detail & Related papers (2023-06-29T17:59:57Z) - Dataless Knowledge Fusion by Merging Weights of Language Models [51.8162883997512]
Fine-tuning pre-trained language models has become the prevalent paradigm for building downstream NLP models.
This creates a barrier to fusing knowledge across individual models to yield a better single model.
We propose a dataless knowledge fusion method that merges models in their parameter space.
arXiv Detail & Related papers (2022-12-19T20:46:43Z) - Data Summarization via Bilevel Optimization [48.89977988203108]
A simple yet powerful approach is to operate on small subsets of data.
In this work, we propose a generic coreset framework that formulates the coreset selection as a cardinality-constrained bilevel optimization problem.
arXiv Detail & Related papers (2021-09-26T09:08:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.