Data-Efficient Deep Learning Method for Image Classification Using Data
Augmentation, Focal Cosine Loss, and Ensemble
- URL: http://arxiv.org/abs/2007.07805v1
- Date: Wed, 15 Jul 2020 16:30:57 GMT
- Title: Data-Efficient Deep Learning Method for Image Classification Using Data
Augmentation, Focal Cosine Loss, and Ensemble
- Authors: Byeongjo Kim, Chanran Kim, Jaehoon Lee, Jein Song, Gyoungsoo Park
- Abstract summary: It is important to leverage small dataset effectively for achieving the better performance.
With these methods, we obtain high accuracy by leveraging ImageNet data which consist of only 50 images per class.
Our model is ranked 4th in Visual Inductive Printers for Data-Effective Computer Vision Challenge.
- Score: 9.55617552923003
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In general, sufficient data is essential for the better performance and
generalization of deep-learning models. However, lots of limitations(cost,
resources, etc.) of data collection leads to lack of enough data in most of the
areas. In addition, various domains of each data sources and licenses also lead
to difficulties in collection of sufficient data. This situation makes us hard
to utilize not only the pre-trained model, but also the external knowledge.
Therefore, it is important to leverage small dataset effectively for achieving
the better performance. We applied some techniques in three aspects: data, loss
function, and prediction to enable training from scratch with less data. With
these methods, we obtain high accuracy by leveraging ImageNet data which
consist of only 50 images per class. Furthermore, our model is ranked 4th in
Visual Inductive Printers for Data-Effective Computer Vision Challenge.
Related papers
- Research and Implementation of Data Enhancement Techniques for Graph Neural Networks [10.575426305555538]
In practical engineering applications, some data are affected by the conditions under which more data cannot be obtained or the cost of obtaining data is too high.
This paper firstly analyses the key points of the data enhancement technology of graph neural network, and at the same time introduces the composition of graph neural network in depth.
arXiv Detail & Related papers (2024-06-18T14:07:38Z) - A Simple and Efficient Baseline for Data Attribution on Images [107.12337511216228]
Current state-of-the-art approaches require a large ensemble of as many as 300,000 models to accurately attribute model predictions.
In this work, we focus on a minimalist baseline, utilizing the feature space of a backbone pretrained via self-supervised learning to perform data attribution.
Our method is model-agnostic and scales easily to large datasets.
arXiv Detail & Related papers (2023-11-03T17:29:46Z) - Data Filtering Networks [67.827994353269]
We study the problem of learning a data filtering network (DFN) for this second step of filtering a large uncurated dataset.
Our key finding is that the quality of a network for filtering is distinct from its performance on downstream tasks.
Based on our insights, we construct new data filtering networks that induce state-of-the-art image-text datasets.
arXiv Detail & Related papers (2023-09-29T17:37:29Z) - Synthetic Data for Object Classification in Industrial Applications [53.180678723280145]
In object classification, capturing a large number of images per object and in different conditions is not always possible.
This work explores the creation of artificial images using a game engine to cope with limited data in the training dataset.
arXiv Detail & Related papers (2022-12-09T11:43:04Z) - Image Data Augmentation for Deep Learning: A Survey [8.817690876855728]
We systematically review different image data augmentation methods.
We propose a taxonomy of reviewed methods and present the strengths and limitations of these methods.
We also conduct extensive experiments with various data augmentation methods on three typical computer vision tasks.
arXiv Detail & Related papers (2022-04-19T02:05:56Z) - On The State of Data In Computer Vision: Human Annotations Remain
Indispensable for Developing Deep Learning Models [0.0]
High-quality labeled datasets play a crucial role in fueling the development of machine learning (ML)
Since the emergence of the ImageNet dataset and the AlexNet model in 2012, the size of new open-source labeled vision datasets has remained roughly constant.
Only a minority of publications in the computer vision community tackle supervised learning on datasets that are orders of magnitude larger than Imagenet.
arXiv Detail & Related papers (2021-07-31T00:08:21Z) - On the Pitfalls of Learning with Limited Data: A Facial Expression
Recognition Case Study [0.5249805590164901]
We focus on the problem of Facial Expression Recognition from videos.
We performed an extensive study with four databases at a different complexity and nine deep-learning architectures for video classification.
We found that complex training sets translate better to more stable test sets when trained with transfer learning and synthetically generated data.
arXiv Detail & Related papers (2021-04-02T18:53:41Z) - Data Impressions: Mining Deep Models to Extract Samples for Data-free
Applications [26.48630545028405]
"Data Impressions" act as proxy to the training data and can be used to realize a variety of tasks.
We show the applicability of data impressions in solving several computer vision tasks.
arXiv Detail & Related papers (2021-01-15T11:37:29Z) - Efficient Conditional Pre-training for Transfer Learning [71.01129334495553]
We propose efficient filtering methods to select relevant subsets from the pre-training dataset.
We validate our techniques by pre-training on ImageNet in both the unsupervised and supervised settings.
We improve standard ImageNet pre-training by 1-3% by tuning available models on our subsets and pre-training on a dataset filtered from a larger scale dataset.
arXiv Detail & Related papers (2020-11-20T06:16:15Z) - Neural Data Server: A Large-Scale Search Engine for Transfer Learning
Data [78.74367441804183]
We introduce Neural Data Server (NDS), a large-scale search engine for finding the most useful transfer learning data to the target domain.
NDS consists of a dataserver which indexes several large popular image datasets, and aims to recommend data to a client.
We show the effectiveness of NDS in various transfer learning scenarios, demonstrating state-of-the-art performance on several target datasets.
arXiv Detail & Related papers (2020-01-09T01:21:30Z) - DeGAN : Data-Enriching GAN for Retrieving Representative Samples from a
Trained Classifier [58.979104709647295]
We bridge the gap between the abundance of available data and lack of relevant data, for the future learning tasks of a trained network.
We use the available data, that may be an imbalanced subset of the original training dataset, or a related domain dataset, to retrieve representative samples.
We demonstrate that data from a related domain can be leveraged to achieve state-of-the-art performance.
arXiv Detail & Related papers (2019-12-27T02:05:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.