Geo-Spatiotemporal Features and Shape-Based Prior Knowledge for
Fine-grained Imbalanced Data Classification
- URL: http://arxiv.org/abs/2103.11285v1
- Date: Sun, 21 Mar 2021 02:01:38 GMT
- Title: Geo-Spatiotemporal Features and Shape-Based Prior Knowledge for
Fine-grained Imbalanced Data Classification
- Authors: Charles (A.) Kantor, Marta Skreta, Brice Rauby, L\'eonard Boussioux,
Emmanuel Jehanno, Alexandra Luccioni, David Rolnick, Hugues Talbot
- Abstract summary: Fine-grained classification aims at distinguishing between items with similar global perception and patterns, but that differ by minute details.
Our primary challenges come from both small inter-class variations and large intra-class variations.
We propose to combine several innovations to improve fine-grained classification within the use-case of wildlife.
- Score: 63.916371837696396
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Fine-grained classification aims at distinguishing between items with similar
global perception and patterns, but that differ by minute details. Our primary
challenges come from both small inter-class variations and large intra-class
variations. In this article, we propose to combine several innovations to
improve fine-grained classification within the use-case of wildlife, which is
of practical interest for experts. We utilize geo-spatiotemporal data to enrich
the picture information and further improve the performance. We also
investigate state-of-the-art methods for handling the imbalanced data issue.
Related papers
- Granularity Matters in Long-Tail Learning [62.30734737735273]
We offer a novel perspective on long-tail learning, inspired by an observation: datasets with finer granularity tend to be less affected by data imbalance.
We introduce open-set auxiliary classes that are visually similar to existing ones, aiming to enhance representation learning for both head and tail classes.
To prevent the overwhelming presence of auxiliary classes from disrupting training, we introduce a neighbor-silencing loss.
arXiv Detail & Related papers (2024-10-21T13:06:21Z) - Localized Gaussians as Self-Attention Weights for Point Clouds Correspondence [92.07601770031236]
We investigate semantically meaningful patterns in the attention heads of an encoder-only Transformer architecture.
We find that fixing the attention weights not only accelerates the training process but also enhances the stability of the optimization.
arXiv Detail & Related papers (2024-09-20T07:41:47Z) - ST-SACLF: Style Transfer Informed Self-Attention Classifier for Bias-Aware Painting Classification [9.534646914709018]
Painting classification plays a vital role in organizing, finding, and suggesting artwork for digital and classic art galleries.
Existing methods struggle with adapting knowledge from the real world to artistic images during training, leading to poor performance when dealing with different datasets.
We generate more data using Style Transfer with Adaptive Instance Normalization (AdaIN), bridging the gap between diverse styles.
We achieve an impressive 87.24% accuracy using the ResNet-50 backbone over 40 training epochs.
arXiv Detail & Related papers (2024-08-03T17:31:58Z) - Geospatial Disparities: A Case Study on Real Estate Prices in Paris [0.3495246564946556]
We propose a toolkit for identifying and mitigating biases arising from geospatial data.
We incorporate an ordinal regression case with spatial attributes, deviating from the binary classification focus.
Illustrating our methodology, we showcase practical applications and scrutinize the implications of choosing geographical aggregation levels for fairness and calibration measures.
arXiv Detail & Related papers (2024-01-29T14:53:14Z) - Fine-grained Recognition with Learnable Semantic Data Augmentation [68.48892326854494]
Fine-grained image recognition is a longstanding computer vision challenge.
We propose diversifying the training data at the feature-level to alleviate the discriminative region loss problem.
Our method significantly improves the generalization performance on several popular classification networks.
arXiv Detail & Related papers (2023-09-01T11:15:50Z) - A Novel Cross-Perturbation for Single Domain Generalization [54.612933105967606]
Single domain generalization aims to enhance the ability of the model to generalize to unknown domains when trained on a single source domain.
The limited diversity in the training data hampers the learning of domain-invariant features, resulting in compromised generalization performance.
We propose CPerb, a simple yet effective cross-perturbation method to enhance the diversity of the training data.
arXiv Detail & Related papers (2023-08-02T03:16:12Z) - Learning from Partially Overlapping Labels: Image Segmentation under
Annotation Shift [68.6874404805223]
We propose several strategies for learning from partially overlapping labels in the context of abdominal organ segmentation.
We find that combining a semi-supervised approach with an adaptive cross entropy loss can successfully exploit heterogeneously annotated data.
arXiv Detail & Related papers (2021-07-13T09:22:24Z) - Novel Visual Category Discovery with Dual Ranking Statistics and Mutual
Knowledge Distillation [16.357091285395285]
We tackle the problem of grouping unlabelled images from new classes into different semantic partitions.
This is a more realistic and challenging setting than conventional semi-supervised learning.
We propose a two-branch learning framework for this problem, with one branch focusing on local part-level information and the other branch focusing on overall characteristics.
arXiv Detail & Related papers (2021-07-07T17:14:40Z) - Neural Networks as Functional Classifiers [0.0]
We extend notable deep learning methodologies to the domain of functional data for the purpose of classification problems.
We highlight the effectiveness of our method in a number of classification applications such as classification of spectrographic data.
arXiv Detail & Related papers (2020-10-09T00:11:01Z) - Learning Cross-domain Generalizable Features by Representation
Disentanglement [11.74643883335152]
Deep learning models exhibit limited generalizability across different domains.
We propose Mutual-Information-based Disentangled Neural Networks (MIDNet) to extract generalizable features that enable transferring knowledge to unseen categorical features in target domains.
We demonstrate our method on handwritten digits datasets and a fetal ultrasound dataset for image classification tasks.
arXiv Detail & Related papers (2020-02-29T17:53:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.