Filtering Empty Camera Trap Images in Embedded Systems
- URL: http://arxiv.org/abs/2104.08859v1
- Date: Sun, 18 Apr 2021 13:56:22 GMT
- Title: Filtering Empty Camera Trap Images in Embedded Systems
- Authors: Fagner Cunha, Eulanda M. dos Santos, Raimundo Barreto, Juan G. Colonna
- Abstract summary: We present a comparative study on animal recognition models to analyze the trade-off between precision and inference latency on edge devices.
The experiments show that, when using the same set of images for training, detectors achieve superior performance.
Considering the high cost of generating labels for the detection problem, when there is a massive number of images labeled for classification, classifiers are able to reach results comparable to detectors but with half latency.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Monitoring wildlife through camera traps produces a massive amount of images,
whose a significant portion does not contain animals, being later discarded.
Embedding deep learning models to identify animals and filter these images
directly in those devices brings advantages such as savings in the storage and
transmission of data, usually resource-constrained in this type of equipment.
In this work, we present a comparative study on animal recognition models to
analyze the trade-off between precision and inference latency on edge devices.
To accomplish this objective, we investigate classifiers and object detectors
of various input resolutions and optimize them using quantization and reducing
the number of model filters. The confidence threshold of each model was
adjusted to obtain 96% recall for the nonempty class, since instances from the
empty class are expected to be discarded. The experiments show that, when using
the same set of images for training, detectors achieve superior performance,
eliminating at least 10% more empty images than classifiers with comparable
latencies. Considering the high cost of generating labels for the detection
problem, when there is a massive number of images labeled for classification
(about one million instances, ten times more than those available for
detection), classifiers are able to reach results comparable to detectors but
with half latency.
Related papers
- Improved detection of discarded fish species through BoxAL active learning [0.2544632696242629]
In this study, we present an active learning technique, named BoxAL, which includes estimation of epistemic certainty of the Faster R-CNN object-detection model.
The method allows selecting the most uncertain training images from an unlabeled pool, which are then used to train the object-detection model.
Our study additionally showed that the sampled new data is more valuable for training than the remaining unlabeled data.
arXiv Detail & Related papers (2024-10-07T10:01:30Z) - Small Effect Sizes in Malware Detection? Make Harder Train/Test Splits! [51.668411293817464]
Industry practitioners care about small improvements in malware detection accuracy because their models are deployed to hundreds of millions of machines.
Academic research is often restrained to public datasets on the order of ten thousand samples.
We devise an approach to generate a benchmark of difficulty from a pool of available samples.
arXiv Detail & Related papers (2023-12-25T21:25:55Z) - Extended target tracking utilizing machine-learning software -- with
applications to animal classification [1.5516470851450592]
This paper considers the problem of detecting and tracking objects in a sequence of images.
The problem is formulated in a filtering framework, using the output of object-detection algorithms as measurements.
An extension to the filtering framework is proposed that incorporates class information from the previous frame to robustify the classification.
arXiv Detail & Related papers (2023-10-12T13:27:21Z) - DISCount: Counting in Large Image Collections with Detector-Based
Importance Sampling [32.522579550452484]
DISCount is a detector-based importance sampling framework for counting in large image collections.
It integrates an imperfect detector with human-in-the-loop screening to produce unbiased estimates of counts.
arXiv Detail & Related papers (2023-06-05T18:04:57Z) - Learning to Annotate Part Segmentation with Gradient Matching [58.100715754135685]
This paper focuses on tackling semi-supervised part segmentation tasks by generating high-quality images with a pre-trained GAN.
In particular, we formulate the annotator learning as a learning-to-learn problem.
We show that our method can learn annotators from a broad range of labelled images including real images, generated images, and even analytically rendered images.
arXiv Detail & Related papers (2022-11-06T01:29:22Z) - Semi-supervised Object Detection via Virtual Category Learning [68.26956850996976]
This paper proposes to use confusing samples proactively without label correction.
Specifically, a virtual category (VC) is assigned to each confusing sample.
It is attributed to specifying the embedding distance between the training sample and the virtual category.
arXiv Detail & Related papers (2022-07-07T16:59:53Z) - Robust and Accurate Object Detection via Adversarial Learning [111.36192453882195]
This work augments the fine-tuning stage for object detectors by exploring adversarial examples.
Our approach boosts the performance of state-of-the-art EfficientDets by +1.1 mAP on the object detection benchmark.
arXiv Detail & Related papers (2021-03-23T19:45:26Z) - Instance Localization for Self-supervised Detection Pretraining [68.24102560821623]
We propose a new self-supervised pretext task, called instance localization.
We show that integration of bounding boxes into pretraining promotes better task alignment and architecture alignment for transfer learning.
Experimental results demonstrate that our approach yields state-of-the-art transfer learning results for object detection.
arXiv Detail & Related papers (2021-02-16T17:58:57Z) - Attention-Aware Noisy Label Learning for Image Classification [97.26664962498887]
Deep convolutional neural networks (CNNs) learned on large-scale labeled samples have achieved remarkable progress in computer vision.
The cheapest way to obtain a large body of labeled visual data is to crawl from websites with user-supplied labels, such as Flickr.
This paper proposes the attention-aware noisy label learning approach to improve the discriminative capability of the network trained on datasets with potential label noise.
arXiv Detail & Related papers (2020-09-30T15:45:36Z) - Multi-species Seagrass Detection and Classification from Underwater
Images [1.2233362977312945]
In this paper, we introduce a multi-species detector and classifier for seagrasses based on a deep convolutional neural network.
We also introduce a simple method to semi-automatically label image patches and therefore minimize manual labelling requirement.
We describe and release publicly the dataset collected in this study as well as the code and pre-trained models to replicate our experiments.
arXiv Detail & Related papers (2020-09-18T07:20:44Z) - Automatic Detection and Recognition of Individuals in Patterned Species [4.163860911052052]
We develop a framework for automatic detection and recognition of individuals in different patterned species.
We use the recently proposed Faster-RCNN object detection framework to efficiently detect animals in images.
We evaluate our recognition system on zebra and jaguar images to show generalization to other patterned species.
arXiv Detail & Related papers (2020-05-06T15:29:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.