Comparison Knowledge Translation for Generalizable Image Classification
- URL: http://arxiv.org/abs/2205.03633v1
- Date: Sat, 7 May 2022 11:05:18 GMT
- Title: Comparison Knowledge Translation for Generalizable Image Classification
- Authors: Zunlei Feng, Tian Qiu, Sai Wu, Xiaotuan Jin, Zengliang He, Mingli
Song, Huiqiong Wang
- Abstract summary: We build a generalizable framework that emulates the humans' recognition mechanism in the image classification task.
We put forward a Comparison Classification Translation Network (CCT-Net), which comprises a comparison classifier and a matching discriminator.
CCT-Net achieves surprising generalization ability on unseen categories and SOTA performance on target categories.
- Score: 31.530232003512957
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep learning has recently achieved remarkable performance in image
classification tasks, which depends heavily on massive annotation. However, the
classification mechanism of existing deep learning models seems to contrast to
humans' recognition mechanism. With only a glance at an image of the object
even unknown type, humans can quickly and precisely find other same category
objects from massive images, which benefits from daily recognition of various
objects. In this paper, we attempt to build a generalizable framework that
emulates the humans' recognition mechanism in the image classification task,
hoping to improve the classification performance on unseen categories with the
support of annotations of other categories. Specifically, we investigate a new
task termed Comparison Knowledge Translation (CKT). Given a set of fully
labeled categories, CKT aims to translate the comparison knowledge learned from
the labeled categories to a set of novel categories. To this end, we put
forward a Comparison Classification Translation Network (CCT-Net), which
comprises a comparison classifier and a matching discriminator. The comparison
classifier is devised to classify whether two images belong to the same
category or not, while the matching discriminator works together in an
adversarial manner to ensure whether classified results match the truth.
Exhaustive experiments show that CCT-Net achieves surprising generalization
ability on unseen categories and SOTA performance on target categories.
Related papers
- Category-Extensible Out-of-Distribution Detection via Hierarchical Context Descriptions [35.20091752343433]
This work introduces two hierarchical contexts, namely perceptual context and spurious context, to carefully describe the precise category boundary.
The two contexts hierarchically construct the precise description for a certain category, which is first roughly classifying a sample to the predicted category.
The precise descriptions for those categories within the vision-language framework present a novel application: CATegory-EXtensible OOD detection (CATEX)
arXiv Detail & Related papers (2024-07-23T12:53:38Z) - ChatGPT-Powered Hierarchical Comparisons for Image Classification [12.126353699873281]
We present a novel framework for image classification based on large language models (LLMs)
We group classes into hierarchies and classifying images by comparing image-text embeddings at each hierarchy level, resulting in an intuitive, effective, and explainable approach.
arXiv Detail & Related papers (2023-11-01T00:26:40Z) - Category Query Learning for Human-Object Interaction Classification [25.979131884959923]
Unlike most previous HOI methods, we propose a novel and complementary approach called category query learning.
This idea is motivated by an earlier multi-label image classification method, but is for the first time applied for the challenging human-object interaction classification task.
Our method is simple, general and effective. It is validated on three representative HOI baselines and achieves new state-of-the-art results on two benchmarks.
arXiv Detail & Related papers (2023-03-24T13:59:58Z) - Stable Attribute Group Editing for Reliable Few-shot Image Generation [88.59350889410794]
We present an editing-based'' framework Attribute Group Editing (AGE) for reliable few-shot image generation.
We find that class inconsistency is a common problem in GAN-generated images for downstream classification.
We propose to boost the downstream classification performance of SAGE by enhancing the pixel and frequency components.
arXiv Detail & Related papers (2023-02-01T01:51:47Z) - Semantic Representation and Dependency Learning for Multi-Label Image
Recognition [76.52120002993728]
We propose a novel and effective semantic representation and dependency learning (SRDL) framework to learn category-specific semantic representation for each category.
Specifically, we design a category-specific attentional regions (CAR) module to generate channel/spatial-wise attention matrices to guide model.
We also design an object erasing (OE) module to implicitly learn semantic dependency among categories by erasing semantic-aware regions.
arXiv Detail & Related papers (2022-04-08T00:55:15Z) - Multi-Label Image Classification with Contrastive Learning [57.47567461616912]
We show that a direct application of contrastive learning can hardly improve in multi-label cases.
We propose a novel framework for multi-label classification with contrastive learning in a fully supervised setting.
arXiv Detail & Related papers (2021-07-24T15:00:47Z) - Boosting few-shot classification with view-learnable contrastive
learning [19.801016732390064]
We introduce contrastive loss into few-shot classification for learning latent fine-grained structure in the embedding space.
We develop a learning-to-learn algorithm to automatically generate different views of the same image.
arXiv Detail & Related papers (2021-07-20T03:13:33Z) - Category Contrast for Unsupervised Domain Adaptation in Visual Tasks [92.9990560760593]
We propose a novel Category Contrast technique (CaCo) that introduces semantic priors on top of instance discrimination for visual UDA tasks.
CaCo is complementary to existing UDA methods and generalizable to other learning setups such as semi-supervised learning, unsupervised model adaptation, etc.
arXiv Detail & Related papers (2021-06-05T12:51:35Z) - Learning and Evaluating Representations for Deep One-class
Classification [59.095144932794646]
We present a two-stage framework for deep one-class classification.
We first learn self-supervised representations from one-class data, and then build one-class classifiers on learned representations.
In experiments, we demonstrate state-of-the-art performance on visual domain one-class classification benchmarks.
arXiv Detail & Related papers (2020-11-04T23:33:41Z) - Zero-Shot Recognition through Image-Guided Semantic Classification [9.291055558504588]
We present a new embedding-based framework for zero-shot learning (ZSL)
Motivated by the binary relevance method for multi-label classification, we propose to inversely learn the mapping between an image and a semantic classifier.
IGSC is conceptually simple and can be realized by a slight enhancement of an existing deep architecture for classification.
arXiv Detail & Related papers (2020-07-23T06:22:40Z) - I Am Going MAD: Maximum Discrepancy Competition for Comparing
Classifiers Adaptively [135.7695909882746]
We name the MAximum Discrepancy (MAD) competition.
We adaptively sample a small test set from an arbitrarily large corpus of unlabeled images.
Human labeling on the resulting model-dependent image sets reveals the relative performance of the competing classifiers.
arXiv Detail & Related papers (2020-02-25T03:32:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.