Fine-grained Species Recognition with Privileged Pooling: Better Sample
Efficiency Through Supervised Attention
- URL: http://arxiv.org/abs/2003.09168v4
- Date: Fri, 4 Aug 2023 14:53:42 GMT
- Title: Fine-grained Species Recognition with Privileged Pooling: Better Sample
Efficiency Through Supervised Attention
- Authors: Andres C. Rodriguez, Stefano D'Aronco, Konrad Schindler and Jan Dirk
Wegner
- Abstract summary: We propose a scheme for supervised image classification that uses privileged information in the form of keypoint annotations for the training data.
Our main motivation is the recognition of animal species for ecological applications such as biodiversity modelling.
In experiments with three different animal species datasets, we show that deep networks with privileged pooling can use small training sets more efficiently and generalize better.
- Score: 26.136331738529243
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a scheme for supervised image classification that uses privileged
information, in the form of keypoint annotations for the training data, to
learn strong models from small and/or biased training sets. Our main motivation
is the recognition of animal species for ecological applications such as
biodiversity modelling, which is challenging because of long-tailed species
distributions due to rare species, and strong dataset biases such as repetitive
scene background in camera traps. To counteract these challenges, we propose a
visual attention mechanism that is supervised via keypoint annotations that
highlight important object parts. This privileged information, implemented as a
novel privileged pooling operation, is only required during training and helps
the model to focus on regions that are discriminative. In experiments with
three different animal species datasets, we show that deep networks with
privileged pooling can use small training sets more efficiently and generalize
better.
Related papers
- Granularity Matters in Long-Tail Learning [62.30734737735273]
We offer a novel perspective on long-tail learning, inspired by an observation: datasets with finer granularity tend to be less affected by data imbalance.
We introduce open-set auxiliary classes that are visually similar to existing ones, aiming to enhance representation learning for both head and tail classes.
To prevent the overwhelming presence of auxiliary classes from disrupting training, we introduce a neighbor-silencing loss.
arXiv Detail & Related papers (2024-10-21T13:06:21Z) - Detecting Endangered Marine Species in Autonomous Underwater Vehicle Imagery Using Point Annotations and Few-Shot Learning [5.439798554380394]
Seafloor imagery collected by Autonomous Underwater Vehicles (AUVs) can be used to identify individuals within their broader habitat context.
Machine learning models can be used to identify the presence of a particular species in images using a trained object detector.
In this paper, inspired by recent work in few-shot learning, images and annotations of common marine species are exploited to enhance the ability of the detector to identify rare and cryptic species.
arXiv Detail & Related papers (2024-06-04T03:31:42Z) - Understanding the Impact of Training Set Size on Animal Re-identification [36.37275024049744]
We show that species-specific characteristics, particularly intra-individual variance, have a notable effect on training data requirements.
We demonstrate the benefits of both local feature and end-to-end learning-based approaches.
arXiv Detail & Related papers (2024-05-24T23:15:52Z) - Towards Seamless Adaptation of Pre-trained Models for Visual Place Recognition [72.35438297011176]
We propose a novel method to realize seamless adaptation of pre-trained models for visual place recognition (VPR)
Specifically, to obtain both global and local features that focus on salient landmarks for discriminating places, we design a hybrid adaptation method.
Experimental results show that our method outperforms the state-of-the-art methods with less training data and training time.
arXiv Detail & Related papers (2024-02-22T12:55:01Z) - Multimodal Foundation Models for Zero-shot Animal Species Recognition in
Camera Trap Images [57.96659470133514]
Motion-activated camera traps constitute an efficient tool for tracking and monitoring wildlife populations across the globe.
Supervised learning techniques have been successfully deployed to analyze such imagery, however training such techniques requires annotations from experts.
Reducing the reliance on costly labelled data has immense potential in developing large-scale wildlife tracking solutions with markedly less human labor.
arXiv Detail & Related papers (2023-11-02T08:32:00Z) - CHALLENGER: Training with Attribution Maps [63.736435657236505]
We show that utilizing attribution maps for training neural networks can improve regularization of models and thus increase performance.
In particular, we show that our generic domain-independent approach yields state-of-the-art results in vision, natural language processing and on time series tasks.
arXiv Detail & Related papers (2022-05-30T13:34:46Z) - SuperAnimal pretrained pose estimation models for behavioral analysis [42.206265576708255]
Quantification of behavior is critical in applications ranging from neuroscience, veterinary medicine and animal conservation efforts.
We present a series of technical innovations that enable a new method, collectively called SuperAnimal, to develop unified foundation models.
arXiv Detail & Related papers (2022-03-14T18:46:57Z) - Deep learning with self-supervision and uncertainty regularization to
count fish in underwater images [28.261323753321328]
Effective conservation actions require effective population monitoring.
Monitoring populations through image sampling has made data collection cheaper, wide-reaching and less intrusive.
Counting animals from such data is challenging, particularly when densely packed in noisy images.
Deep learning is the state-of-the-art method for many computer vision tasks, but it has yet to be properly explored to count animals.
arXiv Detail & Related papers (2021-04-30T13:02:19Z) - Adversarially-Trained Deep Nets Transfer Better: Illustration on Image
Classification [53.735029033681435]
Transfer learning is a powerful methodology for adapting pre-trained deep neural networks on image recognition tasks to new domains.
In this work, we demonstrate that adversarially-trained models transfer better than non-adversarially-trained models.
arXiv Detail & Related papers (2020-07-11T22:48:42Z) - Transferring Dense Pose to Proximal Animal Classes [83.84439508978126]
We show that it is possible to transfer the knowledge existing in dense pose recognition for humans, as well as in more general object detectors and segmenters, to the problem of dense pose recognition in other classes.
We do this by establishing a DensePose model for the new animal which is also geometrically aligned to humans.
We also introduce two benchmark datasets labelled in the manner of DensePose for the class chimpanzee and use them to evaluate our approach.
arXiv Detail & Related papers (2020-02-28T21:43:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.