Your "Flamingo" is My "Bird": Fine-Grained, or Not
- URL: http://arxiv.org/abs/2011.09040v3
- Date: Sun, 28 Mar 2021 06:17:51 GMT
- Title: Your "Flamingo" is My "Bird": Fine-Grained, or Not
- Authors: Dongliang Chang, Kaiyue Pang, Yixiao Zheng, Zhanyu Ma, Yi-Zhe Song,
and Jun Guo
- Abstract summary: We investigate how to tailor for different fine-grained definitions under divergent levels of expertise.
We first conduct a comprehensive human study where we confirm that most participants prefer multi-granularity labels.
We then discover the key intuition that: coarse-level label prediction exacerbates fine-grained feature learning.
- Score: 60.25769809922673
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Whether what you see in Figure 1 is a "flamingo" or a "bird", is the question
we ask in this paper. While fine-grained visual classification (FGVC) strives
to arrive at the former, for the majority of us non-experts just "bird" would
probably suffice. The real question is therefore -- how can we tailor for
different fine-grained definitions under divergent levels of expertise. For
that, we re-envisage the traditional setting of FGVC, from single-label
classification, to that of top-down traversal of a pre-defined coarse-to-fine
label hierarchy -- so that our answer becomes
"bird"-->"Phoenicopteriformes"-->"Phoenicopteridae"-->"flamingo". To approach
this new problem, we first conduct a comprehensive human study where we confirm
that most participants prefer multi-granularity labels, regardless whether they
consider themselves experts. We then discover the key intuition that:
coarse-level label prediction exacerbates fine-grained feature learning, yet
fine-level feature betters the learning of coarse-level classifier. This
discovery enables us to design a very simple albeit surprisingly effective
solution to our new problem, where we (i) leverage level-specific
classification heads to disentangle coarse-level features with fine-grained
ones, and (ii) allow finer-grained features to participate in coarser-grained
label predictions, which in turn helps with better disentanglement. Experiments
show that our method achieves superior performance in the new FGVC setting, and
performs better than state-of-the-art on traditional single-label FGVC problem
as well. Thanks to its simplicity, our method can be easily implemented on top
of any existing FGVC frameworks and is parameter-free.
Related papers
- Active Generalized Category Discovery [60.69060965936214]
Generalized Category Discovery (GCD) endeavors to cluster unlabeled samples from both novel and old classes.
We take the spirit of active learning and propose a new setting called Active Generalized Category Discovery (AGCD)
Our method achieves state-of-the-art performance on both generic and fine-grained datasets.
arXiv Detail & Related papers (2024-03-07T07:12:24Z) - Learning from Label Proportions: Bootstrapping Supervised Learners via Belief Propagation [18.57840057487926]
Learning from Label Proportions (LLP) is a learning problem where only aggregate level labels are available for groups of instances, called bags, during training.
This setting arises in domains like advertising and medicine due to privacy considerations.
We propose a novel algorithmic framework for this problem that iteratively performs two main steps.
arXiv Detail & Related papers (2023-10-12T06:09:26Z) - Shrinking Class Space for Enhanced Certainty in Semi-Supervised Learning [59.44422468242455]
We propose a novel method dubbed ShrinkMatch to learn uncertain samples.
For each uncertain sample, it adaptively seeks a shrunk class space, which merely contains the original top-1 class.
We then impose a consistency regularization between a pair of strongly and weakly augmented samples in the shrunk space to strive for discriminative representations.
arXiv Detail & Related papers (2023-08-13T14:05:24Z) - Test-Time Amendment with a Coarse Classifier for Fine-Grained
Classification [10.719054378755981]
We present a novel approach for Post-Hoc Correction called Hierarchical Ensembles (HiE)
HiE utilizes label hierarchy to improve the performance of fine-grained classification at test-time using the coarse-grained predictions.
Our approach brings notable gains in top-1 accuracy while significantly decreasing the severity of mistakes as training data decreases for the fine-grained classes.
arXiv Detail & Related papers (2023-02-01T10:55:27Z) - Pointly-Supervised Panoptic Segmentation [106.68888377104886]
We propose a new approach to applying point-level annotations for weakly-supervised panoptic segmentation.
Instead of the dense pixel-level labels used by fully supervised methods, point-level labels only provide a single point for each target as supervision.
We formulate the problem in an end-to-end framework by simultaneously generating panoptic pseudo-masks from point-level labels and learning from them.
arXiv Detail & Related papers (2022-10-25T12:03:51Z) - Learning with an Evolving Class Ontology [82.89062737922869]
Lifelong learners must recognize concept that evolve over time.
A common yet underexplored scenario is learning with labels over time that refine/expand old classes.
arXiv Detail & Related papers (2022-10-10T19:58:23Z) - A Novel Plug-in Module for Fine-Grained Visual Classification [0.19336815376402716]
We propose a novel plug-in module that can be integrated to many common backbones to provide strongly discriminative regions.
Experimental results show that the proposed plugin module outperforms state-of-the-art approaches.
arXiv Detail & Related papers (2022-02-08T12:35:58Z) - Clue Me In: Semi-Supervised FGVC with Out-of-Distribution Data [44.90231337626545]
We propose a novel design specifically aimed at making out-of-distribution data work for semi-supervised visual classification.
Our experimental results reveal that (i) the proposed method yields good robustness against out-of-distribution data, and (ii) it can be equipped with prior arts, boosting their performance.
arXiv Detail & Related papers (2021-12-06T07:22:10Z) - Self-Supervised Learning from Semantically Imprecise Data [7.24935792316121]
Learning from imprecise labels such as "animal" or "bird" is an important capability when expertly labeled training data is scarce.
CHILLAX is a recently proposed method to tackle this task.
We extend CHILLAX with a self-supervised scheme using constrained extrapolation to generate pseudo-labels.
arXiv Detail & Related papers (2021-04-22T07:26:14Z) - Revisiting Deep Local Descriptor for Improved Few-Shot Classification [56.74552164206737]
We show how one can improve the quality of embeddings by leveraging textbfDense textbfClassification and textbfAttentive textbfPooling.
We suggest to pool feature maps by applying attentive pooling instead of the widely used global average pooling (GAP) to prepare embeddings for few-shot classification.
arXiv Detail & Related papers (2021-03-30T00:48:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.