Latent Enhancing AutoEncoder for Occluded Image Classification
- URL: http://arxiv.org/abs/2402.06936v1
- Date: Sat, 10 Feb 2024 12:22:31 GMT
- Title: Latent Enhancing AutoEncoder for Occluded Image Classification
- Authors: Ketan Kotwal, Tanay Deshmukh, and Preeti Gopal
- Abstract summary: We introduce LEARN: Latent Enhancing feAture Reconstruction Network.
An auto-encoder based network that can be incorporated into the classification model before its head.
On the OccludedPASCAL3D+ dataset, the proposed LEARN outperforms standard classification models.
- Score: 2.6217304977339473
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Large occlusions result in a significant decline in image classification
accuracy. During inference, diverse types of unseen occlusions introduce
out-of-distribution data to the classification model, leading to accuracy
dropping as low as 50%. As occlusions encompass spatially connected regions,
conventional methods involving feature reconstruction are inadequate for
enhancing classification performance. We introduce LEARN: Latent Enhancing
feAture Reconstruction Network -- An auto-encoder based network that can be
incorporated into the classification model before its classifier head without
modifying the weights of classification model. In addition to reconstruction
and classification losses, training of LEARN effectively combines intra- and
inter-class losses calculated over its latent space -- which lead to
improvement in recovering latent space of occluded data, while preserving its
class-specific discriminative information. On the OccludedPASCAL3D+ dataset,
the proposed LEARN outperforms standard classification models (VGG16 and
ResNet-50) by a large margin and up to 2% over state-of-the-art methods. In
cross-dataset testing, our method improves the average classification accuracy
by more than 5% over the state-of-the-art methods. In every experiment, our
model consistently maintains excellent accuracy on in-distribution data.
Related papers
- Anti-Collapse Loss for Deep Metric Learning Based on Coding Rate Metric [99.19559537966538]
DML aims to learn a discriminative high-dimensional embedding space for downstream tasks like classification, clustering, and retrieval.
To maintain the structure of embedding space and avoid feature collapse, we propose a novel loss function called Anti-Collapse Loss.
Comprehensive experiments on benchmark datasets demonstrate that our proposed method outperforms existing state-of-the-art methods.
arXiv Detail & Related papers (2024-07-03T13:44:20Z) - DiGeo: Discriminative Geometry-Aware Learning for Generalized Few-Shot
Object Detection [39.937724871284665]
Generalized few-shot object detection aims to achieve precise detection on both base classes with abundant annotations and novel classes with limited training data.
Existing approaches enhance few-shot generalization with the sacrifice of base-class performance.
We propose a new training framework, DiGeo, to learn Geometry-aware features of inter-class separation and intra-class compactness.
arXiv Detail & Related papers (2023-03-16T22:37:09Z) - Classification of datasets with imputed missing values: does imputation
quality matter? [2.7646249774183]
Classifying samples in incomplete datasets is non-trivial.
We demonstrate how the commonly used measures for assessing quality are flawed.
We propose a new class of discrepancy scores which focus on how well the method recreates the overall distribution of the data.
arXiv Detail & Related papers (2022-06-16T22:58:03Z) - Exploring Category-correlated Feature for Few-shot Image Classification [27.13708881431794]
We present a simple yet effective feature rectification method by exploring the category correlation between novel and base classes as the prior knowledge.
The proposed approach consistently obtains considerable performance gains on three widely used benchmarks.
arXiv Detail & Related papers (2021-12-14T08:25:24Z) - Calibrating Class Activation Maps for Long-Tailed Visual Recognition [60.77124328049557]
We present two effective modifications of CNNs to improve network learning from long-tailed distribution.
First, we present a Class Activation Map (CAMC) module to improve the learning and prediction of network classifiers.
Second, we investigate the use of normalized classifiers for representation learning in long-tailed problems.
arXiv Detail & Related papers (2021-08-29T05:45:03Z) - Semantic Perturbations with Normalizing Flows for Improved
Generalization [62.998818375912506]
We show that perturbations in the latent space can be used to define fully unsupervised data augmentations.
We find that our latent adversarial perturbations adaptive to the classifier throughout its training are most effective.
arXiv Detail & Related papers (2021-08-18T03:20:00Z) - No Fear of Heterogeneity: Classifier Calibration for Federated Learning
with Non-IID Data [78.69828864672978]
A central challenge in training classification models in the real-world federated system is learning with non-IID data.
We propose a novel and simple algorithm called Virtual Representations (CCVR), which adjusts the classifier using virtual representations sampled from an approximated ssian mixture model.
Experimental results demonstrate that CCVR state-of-the-art performance on popular federated learning benchmarks including CIFAR-10, CIFAR-100, and CINIC-10.
arXiv Detail & Related papers (2021-06-09T12:02:29Z) - Generalized Focal Loss: Learning Qualified and Distributed Bounding
Boxes for Dense Object Detection [85.53263670166304]
One-stage detector basically formulates object detection as dense classification and localization.
Recent trend for one-stage detectors is to introduce an individual prediction branch to estimate the quality of localization.
This paper delves into the representations of the above three fundamental elements: quality estimation, classification and localization.
arXiv Detail & Related papers (2020-06-08T07:24:33Z) - Fine-Grained Visual Classification with Efficient End-to-end
Localization [49.9887676289364]
We present an efficient localization module that can be fused with a classification network in an end-to-end setup.
We evaluate the new model on the three benchmark datasets CUB200-2011, Stanford Cars and FGVC-Aircraft.
arXiv Detail & Related papers (2020-05-11T14:07:06Z) - Discriminative Multi-level Reconstruction under Compact Latent Space for
One-Class Novelty Detection [19.675670045855874]
In one-class novelty detection, a model learns solely on the in-class data to single out out out-class instances.
We propose Discriminative Compact AE that learns both compact and collapse-free latent representations of the in-class data.
arXiv Detail & Related papers (2020-03-03T17:45:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.