Towards Efficient and Data Agnostic Image Classification Training
Pipeline for Embedded Systems
- URL: http://arxiv.org/abs/2108.07049v1
- Date: Mon, 16 Aug 2021 12:38:05 GMT
- Title: Towards Efficient and Data Agnostic Image Classification Training
Pipeline for Embedded Systems
- Authors: Kirill Prokofiev and Vladislav Sovrasov
- Abstract summary: This work is focusing on reviewing the latest augmentation and regularization methods for the image classification.
We can achieve a reasonable performance on a variety of downstream image classification tasks without manual tuning of parameters to each particular task.
Resulting models are computationally efficient and can be deployed to CPU using the OpenVINO toolkit.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Nowadays deep learning-based methods have achieved a remarkable progress at
the image classification task among a wide range of commonly used datasets
(ImageNet, CIFAR, SVHN, Caltech 101, SUN397, etc.). SOTA performance on each of
the mentioned datasets is obtained by careful tuning of the model architecture
and training tricks according to the properties of the target data. Although
this approach allows setting academic records, it is unrealistic that an
average data scientist would have enough resources to build a sophisticated
training pipeline for every image classification task he meets in practice.
This work is focusing on reviewing the latest augmentation and regularization
methods for the image classification and exploring ways to automatically choose
some of the most important hyperparameters: total number of epochs, initial
learning rate value and it's schedule. Having a training procedure equipped
with a lightweight modern CNN architecture (like bileNetV3 or EfficientNet),
sufficient level of regularization and adaptive to data learning rate schedule,
we can achieve a reasonable performance on a variety of downstream image
classification tasks without manual tuning of parameters to each particular
task. Resulting models are computationally efficient and can be deployed to CPU
using the OpenVINO toolkit. Source code is available as a part of the OpenVINO
Training Extensions (https://github.com/openvinotoolkit/training_extensions).
Related papers
- Effective pruning of web-scale datasets based on complexity of concept
clusters [48.125618324485195]
We present a method for pruning large-scale multimodal datasets for training CLIP-style models on ImageNet.
We find that training on a smaller set of high-quality data can lead to higher performance with significantly lower training costs.
We achieve a new state-of-the-art Imagehttps://info.arxiv.org/help/prep#commentsNet zero-shot accuracy and a competitive average zero-shot accuracy on 38 evaluation tasks.
arXiv Detail & Related papers (2024-01-09T14:32:24Z) - A Simple and Efficient Baseline for Data Attribution on Images [107.12337511216228]
Current state-of-the-art approaches require a large ensemble of as many as 300,000 models to accurately attribute model predictions.
In this work, we focus on a minimalist baseline, utilizing the feature space of a backbone pretrained via self-supervised learning to perform data attribution.
Our method is model-agnostic and scales easily to large datasets.
arXiv Detail & Related papers (2023-11-03T17:29:46Z) - Dataset Quantization [72.61936019738076]
We present dataset quantization (DQ), a new framework to compress large-scale datasets into small subsets.
DQ is the first method that can successfully distill large-scale datasets such as ImageNet-1k with a state-of-the-art compression ratio.
arXiv Detail & Related papers (2023-08-21T07:24:29Z) - DINOv2: Learning Robust Visual Features without Supervision [75.42921276202522]
This work shows that existing pretraining methods, especially self-supervised methods, can produce such features if trained on enough curated data from diverse sources.
Most of the technical contributions aim at accelerating and stabilizing the training at scale.
In terms of data, we propose an automatic pipeline to build a dedicated, diverse, and curated image dataset instead of uncurated data, as typically done in the self-supervised literature.
arXiv Detail & Related papers (2023-04-14T15:12:19Z) - How to Train an Accurate and Efficient Object Detection Model on Any
Dataset [0.0]
We propose a dataset-agnostic template for object detection trainings.
It consists of carefully chosen and pre-trained models together with a robust training pipeline for further training.
Our solution works out-of-the-box and provides a strong baseline on a wide range of datasets.
arXiv Detail & Related papers (2022-11-30T17:09:01Z) - EfficientTrain: Exploring Generalized Curriculum Learning for Training
Visual Backbones [80.662250618795]
This paper presents a new curriculum learning approach for the efficient training of visual backbones (e.g., vision Transformers)
As an off-the-shelf method, it reduces the wall-time training cost of a wide variety of popular models by >1.5x on ImageNet-1K/22K without sacrificing accuracy.
arXiv Detail & Related papers (2022-11-17T17:38:55Z) - Pushing the Limits of Simple Pipelines for Few-Shot Learning: External
Data and Fine-Tuning Make a Difference [74.80730361332711]
Few-shot learning is an important and topical problem in computer vision.
We show that a simple transformer-based pipeline yields surprisingly good performance on standard benchmarks.
arXiv Detail & Related papers (2022-04-15T02:55:58Z) - Efficient deep learning models for land cover image classification [0.29748898344267777]
This work experiments with the BigEarthNet dataset for land use land cover (LULC) image classification.
We benchmark different state-of-the-art models, including Convolution Neural Networks, Multi-Layer Perceptrons, Visual Transformers, EfficientNets and Wide Residual Networks (WRN)
Our proposed lightweight model has an order of magnitude less trainable parameters, achieves 4.5% higher averaged f-score classification accuracy for all 19 LULC classes and is trained two times faster with respect to a ResNet50 state-of-the-art model that we use as a baseline.
arXiv Detail & Related papers (2021-11-18T00:03:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.