CRL: Class Representative Learning for Image Classification
- URL: http://arxiv.org/abs/2002.06619v1
- Date: Sun, 16 Feb 2020 17:02:59 GMT
- Title: CRL: Class Representative Learning for Image Classification
- Authors: Mayanka Chandrashekar and Yugyung Lee
- Abstract summary: We propose a novel model, called Class Representative Learning Model (CRL), that can be especially effective in image classification influenced by ZSL.
In the CRL model, first, the learning step is to build class representatives to represent classes in datasets by aggregating prominent features extracted from a Convolutional Neural Network (CNN)
The proposed CRL model demonstrated superior performance compared to the current state-of-the-art research in ZSL and mobile deep learning.
- Score: 5.11566193457943
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Building robust and real-time classifiers with diverse datasets are one of
the most significant challenges to deep learning researchers. It is because
there is a considerable gap between a model built with training (seen) data and
real (unseen) data in applications. Recent works including Zero-Shot Learning
(ZSL), have attempted to deal with this problem of overcoming the apparent gap
through transfer learning. In this paper, we propose a novel model, called
Class Representative Learning Model (CRL), that can be especially effective in
image classification influenced by ZSL. In the CRL model, first, the learning
step is to build class representatives to represent classes in datasets by
aggregating prominent features extracted from a Convolutional Neural Network
(CNN). Second, the inferencing step in CRL is to match between the class
representatives and new data. The proposed CRL model demonstrated superior
performance compared to the current state-of-the-art research in ZSL and mobile
deep learning. The proposed CRL model has been implemented and evaluated in a
parallel environment, using Apache Spark, for both distributed learning and
recognition. An extensive experimental study on the benchmark datasets,
ImageNet-1K, CalTech-101, CalTech-256, CIFAR-100, shows that CRL can build a
class distribution model with drastic improvement in learning and recognition
performance without sacrificing accuracy compared to the state-of-the-art
performances in image classification.
Related papers
- A Closer Look at Benchmarking Self-Supervised Pre-training with Image Classification [51.35500308126506]
Self-supervised learning (SSL) is a machine learning approach where the data itself provides supervision, eliminating the need for external labels.
We study how classification-based evaluation protocols for SSL correlate and how well they predict downstream performance on different dataset types.
arXiv Detail & Related papers (2024-07-16T23:17:36Z) - The Languini Kitchen: Enabling Language Modelling Research at Different
Scales of Compute [66.84421705029624]
We introduce an experimental protocol that enables model comparisons based on equivalent compute, measured in accelerator hours.
We pre-process an existing large, diverse, and high-quality dataset of books that surpasses existing academic benchmarks in quality, diversity, and document length.
This work also provides two baseline models: a feed-forward model derived from the GPT-2 architecture and a recurrent model in the form of a novel LSTM with ten-fold throughput.
arXiv Detail & Related papers (2023-09-20T10:31:17Z) - DiRaC-I: Identifying Diverse and Rare Training Classes for Zero-Shot
Learning [6.75714270653184]
It is intuitive that intelligently selecting the training classes from a dataset for Zero-Shot Learning (ZSL) can improve the performance of existing ZSL methods.
We propose a framework called Diverse and Rare Class Identifier (DiRaC-I) which, given an attribute-based dataset, can intelligently yield the most suitable "seen classes" for training ZSL models.
Our results demonstrate DiRaC-I helps ZSL models to achieve significant classification accuracy improvements.
arXiv Detail & Related papers (2022-12-31T16:05:09Z) - Efficient Gaussian Process Model on Class-Imbalanced Datasets for
Generalized Zero-Shot Learning [37.00463358780726]
We propose a Neural Network model that learns a latent feature embedding and a Gaussian Process (GP) regression model that predicts latent feature prototypes of unseen classes.
Our model is trained efficiently with a simple training strategy that mitigates the impact of class-imbalanced training data.
arXiv Detail & Related papers (2022-10-11T04:57:20Z) - R-DFCIL: Relation-Guided Representation Learning for Data-Free Class
Incremental Learning [64.7996065569457]
Class-Incremental Learning (CIL) struggles with catastrophic forgetting when learning new knowledge.
Recent DFCIL works introduce techniques such as model inversion to synthesize data for previous classes, they fail to overcome forgetting due to the severe domain gap between the synthetic and real data.
This paper proposes relation-guided representation learning (RRL) for DFCIL, dubbed R-DFCIL.
arXiv Detail & Related papers (2022-03-24T14:54:15Z) - Self-Supervised Class Incremental Learning [51.62542103481908]
Existing Class Incremental Learning (CIL) methods are based on a supervised classification framework sensitive to data labels.
When updating them based on the new class data, they suffer from catastrophic forgetting: the model cannot discern old class data clearly from the new.
In this paper, we explore the performance of Self-Supervised representation learning in Class Incremental Learning (SSCIL) for the first time.
arXiv Detail & Related papers (2021-11-18T06:58:19Z) - Calibrating Class Activation Maps for Long-Tailed Visual Recognition [60.77124328049557]
We present two effective modifications of CNNs to improve network learning from long-tailed distribution.
First, we present a Class Activation Map (CAMC) module to improve the learning and prediction of network classifiers.
Second, we investigate the use of normalized classifiers for representation learning in long-tailed problems.
arXiv Detail & Related papers (2021-08-29T05:45:03Z) - No Fear of Heterogeneity: Classifier Calibration for Federated Learning
with Non-IID Data [78.69828864672978]
A central challenge in training classification models in the real-world federated system is learning with non-IID data.
We propose a novel and simple algorithm called Virtual Representations (CCVR), which adjusts the classifier using virtual representations sampled from an approximated ssian mixture model.
Experimental results demonstrate that CCVR state-of-the-art performance on popular federated learning benchmarks including CIFAR-10, CIFAR-100, and CINIC-10.
arXiv Detail & Related papers (2021-06-09T12:02:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.