Watch out Venomous Snake Species: A Solution to SnakeCLEF2023
- URL: http://arxiv.org/abs/2307.09748v1
- Date: Wed, 19 Jul 2023 04:59:58 GMT
- Title: Watch out Venomous Snake Species: A Solution to SnakeCLEF2023
- Authors: Feiran Hu, Peng Wang, Yangyang Li, Chenlong Duan, Zijian Zhu, Fei
Wang, Faen Zhang, Yong Li, Xiu-Shen Wei
- Abstract summary: The SnakeCLEF2023 competition aims to the development of advanced algorithms for snake species identification.
This paper presents a method leveraging utilization of both images and metadata.
Our method achieves 91.31% score of the final metric combined of F1 and other metrics on private leaderboard.
- Score: 27.7177597421459
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The SnakeCLEF2023 competition aims to the development of advanced algorithms
for snake species identification through the analysis of images and
accompanying metadata. This paper presents a method leveraging utilization of
both images and metadata. Modern CNN models and strong data augmentation are
utilized to learn better representation of images. To relieve the challenge of
long-tailed distribution, seesaw loss is utilized in our method. We also design
a light model to calculate prior probabilities using metadata features
extracted from CLIP in post processing stage. Besides, we attach more
importance to venomous species by assigning venomous species labels to some
examples that model is uncertain about. Our method achieves 91.31% score of the
final metric combined of F1 and other metrics on private leaderboard, which is
the 1st place among the participators. The code is available at
https://github.com/xiaoxsparraw/CLEF2023.
Related papers
- Improved detection of discarded fish species through BoxAL active learning [0.2544632696242629]
In this study, we present an active learning technique, named BoxAL, which includes estimation of epistemic certainty of the Faster R-CNN object-detection model.
The method allows selecting the most uncertain training images from an unlabeled pool, which are then used to train the object-detection model.
Our study additionally showed that the sampled new data is more valuable for training than the remaining unlabeled data.
arXiv Detail & Related papers (2024-10-07T10:01:30Z) - Transfer Learning with Self-Supervised Vision Transformers for Snake Identification [0.0]
We present our approach for the SnakeCLEF 2024 competition to predict snake species from images.
We use Meta's DINOv2 vision transformer model for feature extraction to tackle species' high variability and visual similarity in a dataset of 182,261 images.
Despite achieving a score of 39.69, our results show promise for DINOv2 embeddings in snake identification.
arXiv Detail & Related papers (2024-07-08T17:52:23Z) - Data Attribution for Text-to-Image Models by Unlearning Synthesized Images [71.23012718682634]
The goal of data attribution for text-to-image models is to identify the training images that most influence the generation of a new image.
We propose a new approach that efficiently identifies highly-influential images.
arXiv Detail & Related papers (2024-06-13T17:59:44Z) - Raising the Bar of AI-generated Image Detection with CLIP [50.345365081177555]
The aim of this work is to explore the potential of pre-trained vision-language models (VLMs) for universal detection of AI-generated images.
We develop a lightweight detection strategy based on CLIP features and study its performance in a wide variety of challenging scenarios.
arXiv Detail & Related papers (2023-11-30T21:11:20Z) - Solutions for Fine-grained and Long-tailed Snake Species Recognition in
SnakeCLEF 2022 [30.8004334312293]
We introduce our solution in SnakeCLEF 2022 for fine-grained snake species recognition on a heavy long-tailed class distribution.
With an ensemble of several different models, a private score 82.65%, ranking the 3rd, is achieved on the final leaderboard.
arXiv Detail & Related papers (2022-07-04T05:55:58Z) - Understanding Gender and Racial Disparities in Image Recognition Models [0.0]
We investigate a multi-label softmax loss with cross-entropy as the loss function instead of a binary cross-entropy on a multi-label classification problem.
We use the MR2 dataset to evaluate the fairness in the model outcomes and try to interpret the mistakes by looking at model activations and suggest possible fixes.
arXiv Detail & Related papers (2021-07-20T01:05:31Z) - NP-DRAW: A Non-Parametric Structured Latent Variable Modelfor Image
Generation [139.8037697822064]
We present a non-parametric structured latent variable model for image generation, called NP-DRAW.
It sequentially draws on a latent canvas in a part-by-part fashion and then decodes the image from the canvas.
arXiv Detail & Related papers (2021-06-25T05:17:55Z) - An Interaction-based Convolutional Neural Network (ICNN) Towards Better
Understanding of COVID-19 X-ray Images [0.0]
We propose a novel Interaction-based Convolutional Neural Network (ICNN) that does not make assumptions about the relevance of local information.
We demonstrate that the proposed method produces state-of-the-art prediction performance of 99.8% on a real-world data set classifying COVID-19 Chest X-ray images.
arXiv Detail & Related papers (2021-06-13T04:41:17Z) - Towards Good Practices for Efficiently Annotating Large-Scale Image
Classification Datasets [90.61266099147053]
We investigate efficient annotation strategies for collecting multi-class classification labels for a large collection of images.
We propose modifications and best practices aimed at minimizing human labeling effort.
Simulated experiments on a 125k image subset of the ImageNet100 show that it can be annotated to 80% top-1 accuracy with 0.35 annotations per image on average.
arXiv Detail & Related papers (2021-04-26T16:29:32Z) - Metadata Normalization [54.43363251520749]
Batch Normalization (BN) normalizes feature distributions by standardizing with batch statistics.
BN does not correct the influence on features from extraneous variables or multiple distributions.
We introduce the Metadata Normalization layer, a new batch-level operation which can be used end-to-end within the training framework.
arXiv Detail & Related papers (2021-04-19T05:10:26Z) - SCAN: Learning to Classify Images without Labels [73.69513783788622]
We advocate a two-step approach where feature learning and clustering are decoupled.
A self-supervised task from representation learning is employed to obtain semantically meaningful features.
We obtain promising results on ImageNet, and outperform several semi-supervised learning methods in the low-data regime.
arXiv Detail & Related papers (2020-05-25T18:12:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.