Weakly Supervised Faster-RCNN+FPN to classify animals in camera trap
images
- URL: http://arxiv.org/abs/2208.14060v1
- Date: Tue, 30 Aug 2022 08:21:59 GMT
- Title: Weakly Supervised Faster-RCNN+FPN to classify animals in camera trap
images
- Authors: Pierrick Pochelu, Clara Erard, Philippe Cordier, Serge G. Petiton,
Bruno Conche
- Abstract summary: Camera traps have revolutionized the animal research of many species that were previously nearly impossible to observe due to their habitat or behavior.
Deep learning has the potential to overcome the workload to automate image classification according to taxon or empty images.
We propose a workflow named Weakly Object Detection Faster-RCNN+FPN which suits this challenge.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Camera traps have revolutionized the animal research of many species that
were previously nearly impossible to observe due to their habitat or behavior.
They are cameras generally fixed to a tree that take a short sequence of images
when triggered. Deep learning has the potential to overcome the workload to
automate image classification according to taxon or empty images. However, a
standard deep neural network classifier fails because animals often represent a
small portion of the high-definition images. That is why we propose a workflow
named Weakly Object Detection Faster-RCNN+FPN which suits this challenge. The
model is weakly supervised because it requires only the animal taxon label per
image but doesn't require any manual bounding box annotations. First, it
automatically performs the weakly-supervised bounding box annotation using the
motion from multiple frames. Then, it trains a Faster-RCNN+FPN model using this
weak supervision. Experimental results have been obtained with two datasets
from a Papua New Guinea and Missouri biodiversity monitoring campaign, then on
an easily reproducible testbed.
Related papers
- POLO -- Point-based, multi-class animal detection [4.572449721501902]
POLO is a multi-class object detection model that can be trained entirely on point labels.
POLO achieves improved accuracy in counting animals in aerial imagery.
arXiv Detail & Related papers (2024-10-15T16:17:16Z) - PARDINUS: Weakly supervised discarding of photo-trapping empty images
based on autoencoders [1.3654846342364308]
Photo-trapping cameras take photographs when motion is detected to capture images where animals appear.
A significant portion of these images are empty - no wildlife appears in the image.
Filtering out those images is not a trivial task since it requires hours of manual work from biologists.
arXiv Detail & Related papers (2023-12-22T16:33:45Z) - Multimodal Foundation Models for Zero-shot Animal Species Recognition in
Camera Trap Images [57.96659470133514]
Motion-activated camera traps constitute an efficient tool for tracking and monitoring wildlife populations across the globe.
Supervised learning techniques have been successfully deployed to analyze such imagery, however training such techniques requires annotations from experts.
Reducing the reliance on costly labelled data has immense potential in developing large-scale wildlife tracking solutions with markedly less human labor.
arXiv Detail & Related papers (2023-11-02T08:32:00Z) - Inverting the Imaging Process by Learning an Implicit Camera Model [73.81635386829846]
This paper proposes a novel implicit camera model which represents the physical imaging process of a camera as a deep neural network.
We demonstrate the power of this new implicit camera model on two inverse imaging tasks.
arXiv Detail & Related papers (2023-04-25T11:55:03Z) - CamDiff: Camouflage Image Augmentation via Diffusion Model [83.35960536063857]
CamDiff is a novel approach to synthesize salient objects in camouflaged scenes.
We leverage the latent diffusion model to synthesize salient objects in camouflaged scenes.
Our approach enables flexible editing and efficient large-scale dataset generation at a low cost.
arXiv Detail & Related papers (2023-04-11T19:37:47Z) - NEWTON: Neural View-Centric Mapping for On-the-Fly Large-Scale SLAM [51.21564182169607]
Newton is a view-centric mapping method that dynamically constructs neural fields based on run-time observation.
Our method enables camera pose updates using loop closures and scene boundary updates by representing the scene with multiple neural fields.
The experimental results demonstrate the superior performance of our method over existing world-centric neural field-based SLAM systems.
arXiv Detail & Related papers (2023-03-23T20:22:01Z) - Robustifying the Multi-Scale Representation of Neural Radiance Fields [86.69338893753886]
We present a robust multi-scale neural radiance fields representation approach to overcome both real-world imaging issues.
Our method handles multi-scale imaging effects and camera-pose estimation problems with NeRF-inspired approaches.
We demonstrate, with examples, that for an accurate neural representation of an object from day-to-day acquired multi-view images, it is crucial to have precise camera-pose estimates.
arXiv Detail & Related papers (2022-10-09T11:46:45Z) - Choosing an Appropriate Platform and Workflow for Processing Camera Trap
Data using Artificial Intelligence [0.18350044465969417]
Camera traps have transformed how ecologists study wildlife species distributions, activity patterns, and interspecific interactions.
The potential of Artificial Intelligence (AI), specifically Deep Learning (DL), to process camera-trap data has gained considerable attention.
Using DL for these applications involves training algorithms, such as Convolutional Neural Networks (CNNs) to automatically detect objects and classify species.
arXiv Detail & Related papers (2022-02-04T18:13:09Z) - Counterfactual Generative Networks [59.080843365828756]
We propose to decompose the image generation process into independent causal mechanisms that we train without direct supervision.
By exploiting appropriate inductive biases, these mechanisms disentangle object shape, object texture, and background.
We show that the counterfactual images can improve out-of-distribution with a marginal drop in performance on the original classification task.
arXiv Detail & Related papers (2021-01-15T10:23:12Z) - Automatic Detection and Recognition of Individuals in Patterned Species [4.163860911052052]
We develop a framework for automatic detection and recognition of individuals in different patterned species.
We use the recently proposed Faster-RCNN object detection framework to efficiently detect animals in images.
We evaluate our recognition system on zebra and jaguar images to show generalization to other patterned species.
arXiv Detail & Related papers (2020-05-06T15:29:21Z) - Sequence Information Channel Concatenation for Improving Camera Trap
Image Burst Classification [1.94742788320879]
Camera Traps are extensively used to observe wildlife in their natural habitat without disturbing the ecosystem.
Currently, a massive number of such camera traps have been deployed at various ecological conservation areas around the world, collecting data for decades.
Existing systems perform classification to detect if images contain animals by considering a single image.
We show that concatenating masks containing sequence information and the images from the 3-image-burst across channels, improves the ROC AUC by 20% on a test-set from unseen camera-sites.
arXiv Detail & Related papers (2020-04-30T21:47:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.