Enhancing Classification with Hierarchical Scalable Query on Fusion
Transformer
- URL: http://arxiv.org/abs/2302.14487v1
- Date: Tue, 28 Feb 2023 11:00:55 GMT
- Title: Enhancing Classification with Hierarchical Scalable Query on Fusion
Transformer
- Authors: Sudeep Kumar Sahoo, Sathish Chalasani, Abhishek Joshi and Kiran
Nanjunda Iyer
- Abstract summary: This paper proposes a method to boost fine-grained classification through a hierarchical approach via learnable independent query embeddings.
We exploit the idea of hierarchy to learn query embeddings that are scalable across all levels.
Our method is able to outperform the existing methods with an improvement of 11% at the fine-grained classification.
- Score: 0.4129225533930965
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Real-world vision based applications require fine-grained classification for
various area of interest like e-commerce, mobile applications, warehouse
management, etc. where reducing the severity of mistakes and improving the
classification accuracy is of utmost importance. This paper proposes a method
to boost fine-grained classification through a hierarchical approach via
learnable independent query embeddings. This is achieved through a
classification network that uses coarse class predictions to improve the fine
class accuracy in a stage-wise sequential manner. We exploit the idea of
hierarchy to learn query embeddings that are scalable across all levels, thus
making this a relevant approach even for extreme classification where we have a
large number of classes. The query is initialized with a weighted Eigen image
calculated from training samples to best represent and capture the variance of
the object. We introduce transformer blocks to fuse intermediate layers at
which query attention happens to enhance the spatial representation of feature
maps at different scales. This multi-scale fusion helps improve the accuracy of
small-size objects. We propose a two-fold approach for the unique
representation of learnable queries. First, at each hierarchical level, we
leverage cluster based loss that ensures maximum separation between inter-class
query embeddings and helps learn a better (query) representation in higher
dimensional spaces. Second, we fuse coarse level queries with finer level
queries weighted by a learned scale factor. We additionally introduce a novel
block called Cross Attention on Multi-level queries with Prior (CAMP) Block
that helps reduce error propagation from coarse level to finer level, which is
a common problem in all hierarchical classifiers. Our method is able to
outperform the existing methods with an improvement of ~11% at the fine-grained
classification.
Related papers
- Harnessing Superclasses for Learning from Hierarchical Databases [1.835004446596942]
In many large-scale classification problems, classes are organized in a known hierarchy, typically represented as a tree.
We introduce a loss for this type of supervised hierarchical classification.
Our approach does not entail any significant additional computational cost compared with the loss of cross-entropy.
arXiv Detail & Related papers (2024-11-25T14:39:52Z) - Efficient Subclass Segmentation in Medical Images [3.383033695275859]
One feasible way to reduce the cost is to annotate with coarse-grained superclass labels while using limited fine-grained annotations as a complement.
There is a lack of research on efficient learning of fine-grained subclasses in semantic segmentation tasks.
Our approach achieves comparable accuracy to a model trained with full subclass annotations, with limited subclass annotations and sufficient superclass annotations.
arXiv Detail & Related papers (2023-07-01T07:39:08Z) - Bi-directional Feature Reconstruction Network for Fine-Grained Few-Shot
Image Classification [61.411869453639845]
We introduce a bi-reconstruction mechanism that can simultaneously accommodate for inter-class and intra-class variations.
This design effectively helps the model to explore more subtle and discriminative features.
Experimental results on three widely used fine-grained image classification datasets consistently show considerable improvements.
arXiv Detail & Related papers (2022-11-30T16:55:14Z) - Hierarchical classification at multiple operating points [1.520694326234112]
We present an efficient algorithm to produce operating characteristic curves for any method that assigns a score to every class in the hierarchy.
We propose two novel loss functions and show that a soft variant of the structured hinge loss is able to significantly outperform the flat baseline.
arXiv Detail & Related papers (2022-10-19T23:36:16Z) - Decision Making for Hierarchical Multi-label Classification with
Multidimensional Local Precision Rate [4.812468844362369]
We introduce a new statistic called the multidimensional local precision rate (mLPR) for each object in each class.
We show that classification decisions made by simply sorting objects across classes in descending order of their mLPRs can, in theory, ensure the class hierarchy.
In response, we introduce HierRank, a new algorithm that maximizes an empirical version of CATCH using estimated mLPRs while respecting the hierarchy.
arXiv Detail & Related papers (2022-05-16T17:43:35Z) - Multiple Classifiers Based Maximum Classifier Discrepancy for
Unsupervised Domain Adaptation [25.114533037440896]
We propose to extend the structure of two classifiers to multiple classifiers to further boost its performance.
We demonstrate that, on average, adopting the structure of three classifiers normally yields the best performance as a trade-off between the accuracy and efficiency.
arXiv Detail & Related papers (2021-08-02T03:00:13Z) - Revisiting Deep Local Descriptor for Improved Few-Shot Classification [56.74552164206737]
We show how one can improve the quality of embeddings by leveraging textbfDense textbfClassification and textbfAttentive textbfPooling.
We suggest to pool feature maps by applying attentive pooling instead of the widely used global average pooling (GAP) to prepare embeddings for few-shot classification.
arXiv Detail & Related papers (2021-03-30T00:48:28Z) - Binary Classification from Multiple Unlabeled Datasets via Surrogate Set
Classification [94.55805516167369]
We propose a new approach for binary classification from m U-sets for $mge2$.
Our key idea is to consider an auxiliary classification task called surrogate set classification (SSC)
arXiv Detail & Related papers (2021-02-01T07:36:38Z) - Learning and Evaluating Representations for Deep One-class
Classification [59.095144932794646]
We present a two-stage framework for deep one-class classification.
We first learn self-supervised representations from one-class data, and then build one-class classifiers on learned representations.
In experiments, we demonstrate state-of-the-art performance on visual domain one-class classification benchmarks.
arXiv Detail & Related papers (2020-11-04T23:33:41Z) - Fewer is More: A Deep Graph Metric Learning Perspective Using Fewer
Proxies [65.92826041406802]
We propose a Proxy-based deep Graph Metric Learning approach from the perspective of graph classification.
Multiple global proxies are leveraged to collectively approximate the original data points for each class.
We design a novel reverse label propagation algorithm, by which the neighbor relationships are adjusted according to ground-truth labels.
arXiv Detail & Related papers (2020-10-26T14:52:42Z) - Fine-Grained Visual Classification with Efficient End-to-end
Localization [49.9887676289364]
We present an efficient localization module that can be fused with a classification network in an end-to-end setup.
We evaluate the new model on the three benchmark datasets CUB200-2011, Stanford Cars and FGVC-Aircraft.
arXiv Detail & Related papers (2020-05-11T14:07:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.