Sequence Information Channel Concatenation for Improving Camera Trap
Image Burst Classification
- URL: http://arxiv.org/abs/2005.00116v2
- Date: Sat, 6 Jun 2020 02:57:41 GMT
- Title: Sequence Information Channel Concatenation for Improving Camera Trap
Image Burst Classification
- Authors: Bhuvan Malladihalli Shashidhara, Darshan Mehta, Yash Kale, Dan Morris,
Megan Hazen
- Abstract summary: Camera Traps are extensively used to observe wildlife in their natural habitat without disturbing the ecosystem.
Currently, a massive number of such camera traps have been deployed at various ecological conservation areas around the world, collecting data for decades.
Existing systems perform classification to detect if images contain animals by considering a single image.
We show that concatenating masks containing sequence information and the images from the 3-image-burst across channels, improves the ROC AUC by 20% on a test-set from unseen camera-sites.
- Score: 1.94742788320879
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Camera Traps are extensively used to observe wildlife in their natural
habitat without disturbing the ecosystem. This could help in the early
detection of natural or human threats to animals, and help towards ecological
conservation. Currently, a massive number of such camera traps have been
deployed at various ecological conservation areas around the world, collecting
data for decades, thereby requiring automation to detect images containing
animals. Existing systems perform classification to detect if images contain
animals by considering a single image. However, due to challenging scenes with
animals camouflaged in their natural habitat, it sometimes becomes difficult to
identify the presence of animals from merely a single image. We hypothesize
that a short burst of images instead of a single image, assuming that the
animal moves, makes it much easier for a human as well as a machine to detect
the presence of animals. In this work, we explore a variety of approaches, and
measure the impact of using short image sequences (burst of 3 images) on
improving the camera trap image classification. We show that concatenating
masks containing sequence information and the images from the 3-image-burst
across channels, improves the ROC AUC by 20% on a test-set from unseen
camera-sites, as compared to an equivalent model that learns from a single
image.
Related papers
- Multimodal Foundation Models for Zero-shot Animal Species Recognition in
Camera Trap Images [57.96659470133514]
Motion-activated camera traps constitute an efficient tool for tracking and monitoring wildlife populations across the globe.
Supervised learning techniques have been successfully deployed to analyze such imagery, however training such techniques requires annotations from experts.
Reducing the reliance on costly labelled data has immense potential in developing large-scale wildlife tracking solutions with markedly less human labor.
arXiv Detail & Related papers (2023-11-02T08:32:00Z) - Deep Learning for Event-based Vision: A Comprehensive Survey and Benchmarks [55.81577205593956]
Event cameras are bio-inspired sensors that capture the per-pixel intensity changes asynchronously.
Deep learning (DL) has been brought to this emerging field and inspired active research endeavors in mining its potential.
arXiv Detail & Related papers (2023-02-17T14:19:28Z) - Florida Wildlife Camera Trap Dataset [48.99466876948454]
We introduce a challenging wildlife camera trap classification dataset collected from two different locations in Southwestern Florida.
The dataset consists of 104,495 images featuring visually similar species, varying illumination conditions, skewed class distribution, and including samples of endangered species.
arXiv Detail & Related papers (2021-06-23T18:53:15Z) - The iWildCam 2021 Competition Dataset [5.612688040565423]
Ecologists use camera traps to monitor animal populations all over the world.
To estimate the abundance of a species, ecologists need to know not just which species were seen, but how many individuals of each species were seen.
We have prepared a challenge where the training data and test data are from different cameras spread across the globe.
arXiv Detail & Related papers (2021-05-07T20:27:22Z) - An explainable deep vision system for animal classification and
detection in trail-camera images with automatic post-deployment retraining [0.0]
This paper introduces an automated vision system for animal detection in trail-camera images taken from a field under the administration of the Texas Parks and Wildlife Department.
We implement a two-stage deep convolutional neural network pipeline to find animal-containing images in the first stage and then process these images to detect birds in the second stage.
The animal classification system classifies animal images with overall 93% sensitivity and 96% specificity. The bird detection system achieves better than 93% sensitivity, 92% specificity, and 68% average Intersection-over-Union rate.
arXiv Detail & Related papers (2020-10-22T06:29:55Z) - WhoAmI: An Automatic Tool for Visual Recognition of Tiger and Leopard
Individuals in the Wild [3.1708876837195157]
We develop automatic algorithms that are able to detect animals, identify the species of animals and to recognize individual animals for two species.
We demonstrate the effectiveness of our approach on a data set of camera-trap images recorded in the jungles of Southern India.
arXiv Detail & Related papers (2020-06-17T16:17:46Z) - Automatic Detection and Recognition of Individuals in Patterned Species [4.163860911052052]
We develop a framework for automatic detection and recognition of individuals in different patterned species.
We use the recently proposed Faster-RCNN object detection framework to efficiently detect animals in images.
We evaluate our recognition system on zebra and jaguar images to show generalization to other patterned species.
arXiv Detail & Related papers (2020-05-06T15:29:21Z) - VisualEchoes: Spatial Image Representation Learning through Echolocation [97.23789910400387]
Several animal species (e.g., bats, dolphins, and whales) and even visually impaired humans have the remarkable ability to perform echolocation.
We propose a novel interaction-based representation learning framework that learns useful visual features via echolocation.
Our work opens a new path for representation learning for embodied agents, where supervision comes from interacting with the physical world.
arXiv Detail & Related papers (2020-05-04T16:16:58Z) - Transferring Dense Pose to Proximal Animal Classes [83.84439508978126]
We show that it is possible to transfer the knowledge existing in dense pose recognition for humans, as well as in more general object detectors and segmenters, to the problem of dense pose recognition in other classes.
We do this by establishing a DensePose model for the new animal which is also geometrically aligned to humans.
We also introduce two benchmark datasets labelled in the manner of DensePose for the class chimpanzee and use them to evaluate our approach.
arXiv Detail & Related papers (2020-02-28T21:43:53Z) - Automatic image-based identification and biomass estimation of
invertebrates [70.08255822611812]
Time-consuming sorting and identification of taxa pose strong limitations on how many insect samples can be processed.
We propose to replace the standard manual approach of human expert-based sorting and identification with an automatic image-based technology.
We use state-of-the-art Resnet-50 and InceptionV3 CNNs for the classification task.
arXiv Detail & Related papers (2020-02-05T21:38:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.