MMLA: Multi-Environment, Multi-Species, Low-Altitude Aerial Footage Dataset
- URL: http://arxiv.org/abs/2504.07744v1
- Date: Thu, 10 Apr 2025 13:40:27 GMT
- Title: MMLA: Multi-Environment, Multi-Species, Low-Altitude Aerial Footage Dataset
- Authors: Jenna Kline, Samuel Stevens, Guy Maalouf, Camille Rondeau Saint-Jean, Dat Nguyen Ngoc, Majid Mirmehdi, David Guerin, Tilo Burghardt, Elzbieta Pastucha, Blair Costelloe, Matthew Watson, Thomas Richardson, Ulrik Pagh Schultz Lundquist,
- Abstract summary: Real-time wildlife detection in drone imagery is critical for numerous applications, including animal ecology, conservation, and biodiversity monitoring.<n>We present a novel multi-species, multi-environment, low-altitude aerial footage (MMLA) dataset.<n>Results demonstrate significant performance disparities across locations and species-specific detection variations.
- Score: 3.7188931723069443
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Real-time wildlife detection in drone imagery is critical for numerous applications, including animal ecology, conservation, and biodiversity monitoring. Low-altitude drone missions are effective for collecting fine-grained animal movement and behavior data, particularly if missions are automated for increased speed and consistency. However, little work exists on evaluating computer vision models on low-altitude aerial imagery and generalizability across different species and settings. To fill this gap, we present a novel multi-environment, multi-species, low-altitude aerial footage (MMLA) dataset. MMLA consists of drone footage collected across three diverse environments: Ol Pejeta Conservancy and Mpala Research Centre in Kenya, and The Wilds Conservation Center in Ohio, which includes five species: Plains zebras, Grevy's zebras, giraffes, onagers, and African Painted Dogs. We comprehensively evaluate three YOLO models (YOLOv5m, YOLOv8m, and YOLOv11m) for detecting animals. Results demonstrate significant performance disparities across locations and species-specific detection variations. Our work highlights the importance of evaluating detection algorithms across different environments for robust wildlife monitoring applications using drones.
Related papers
- WildLive: Near Real-time Visual Wildlife Tracking onboard UAVs [4.215854427679142]
We introduce WildLive -- a near real-time animal detection and tracking framework for high-resolution imagery running directly onboard aerial vehicles (UAVs)
The system performs multi-animal detection tracking at 17 fps+ for HD and 7 fps+ on 4K video streams suitable for operation during higher altitude flights to minimise animal disturbance.
We introduce our WildLive dataset, which comprises 200k+ annotated animal instances across 19k+ frames from 4K UAV videos collected at the Ol Pejeta Conservancy in Kenya.
arXiv Detail & Related papers (2025-04-14T12:21:16Z) - MammAlps: A multi-view video behavior monitoring dataset of wild mammals in the Swiss Alps [41.58000025132071]
MammAlps is a dataset of wildlife behavior monitoring from 9 camera-traps in the Swiss National Park.<n>Based on 6135 single animal clips, we propose the first hierarchical and multimodal animal behavior recognition benchmark.<n>We also propose a second ecology-oriented benchmark aiming at identifying activities, species, number of individuals and meteorological conditions.
arXiv Detail & Related papers (2025-03-23T21:51:58Z) - Few-shot Species Range Estimation [61.60698161072356]
Knowing where a particular species can or cannot be found on Earth is crucial for ecological research and conservation efforts.<n>We outline a new approach for few-shot species range estimation to address the challenge of accurately estimating the range of a species from limited data.<n>During inference, our model takes a set of spatial locations as input, along with optional metadata such as text or an image, and outputs a species encoding that can be used to predict the range of a previously unseen species in feed-forward manner.
arXiv Detail & Related papers (2025-02-20T19:13:29Z) - BuckTales : A multi-UAV dataset for multi-object tracking and re-identification of wild antelopes [0.6267336085190178]
BuckTales is the first large-scale UAV dataset designed to solve multi-object tracking and re-identification problem in wild animals.
The MOT dataset includes over 1.2 million annotations including 680 tracks across 12 high-resolution (5.4K) videos.
The Re-ID dataset includes 730 individuals captured with two UAVs simultaneously.
arXiv Detail & Related papers (2024-11-11T11:55:14Z) - Multi-Species Object Detection in Drone Imagery for Population Monitoring of Endangered Animals [0.0]
This research focused on fine-tuning object detection models for drone images to create accurate counts of animal species.
We trained 30 different models, with the largest having 43.7 million parameters and 365 layers.
We deployed the models on the Jetson Orin Nano for demonstration of low-power real-time species detection for easy inference on drones.
arXiv Detail & Related papers (2024-06-28T03:03:55Z) - Multiview Aerial Visual Recognition (MAVREC): Can Multi-view Improve
Aerial Visual Perception? [57.77643186237265]
We present Multiview Aerial Visual RECognition or MAVREC, a video dataset where we record synchronized scenes from different perspectives.
MAVREC consists of around 2.5 hours of industry-standard 2.7K resolution video sequences, more than 0.5 million frames, and 1.1 million annotated bounding boxes.
This makes MAVREC the largest ground and aerial-view dataset, and the fourth largest among all drone-based datasets.
arXiv Detail & Related papers (2023-12-07T18:59:14Z) - Multimodal Foundation Models for Zero-shot Animal Species Recognition in
Camera Trap Images [57.96659470133514]
Motion-activated camera traps constitute an efficient tool for tracking and monitoring wildlife populations across the globe.
Supervised learning techniques have been successfully deployed to analyze such imagery, however training such techniques requires annotations from experts.
Reducing the reliance on costly labelled data has immense potential in developing large-scale wildlife tracking solutions with markedly less human labor.
arXiv Detail & Related papers (2023-11-02T08:32:00Z) - SatBird: Bird Species Distribution Modeling with Remote Sensing and
Citizen Science Data [68.2366021016172]
We present SatBird, a satellite dataset of locations in the USA with labels derived from presence-absence observation data from the citizen science database eBird.
We also provide a dataset in Kenya representing low-data regimes.
We benchmark a set of baselines on our dataset, including SOTA models for remote sensing tasks.
arXiv Detail & Related papers (2023-11-02T02:00:27Z) - SOCRATES: A Stereo Camera Trap for Monitoring of Biodiversity [0.0]
This study presents a novel approach to 3D camera trapping featuring highly optimized hardware and software.
A comprehensive evaluation of SOCRATES shows not only a $3.23%$ improvement in animal detection but also its superior applicability for estimating animal abundance.
arXiv Detail & Related papers (2022-09-19T15:03:35Z) - LudVision -- Remote Detection of Exotic Invasive Aquatic Floral Species
using Drone-Mounted Multispectral Data [0.34998703934432673]
Ludwigia peploides is considered by the European Union as an aquatic invasive species.
Our goal was to develop a method to identify the presence of the species.
To identify the targeted species on the collected images, we propose a new method for detecting Ludwigia p.
arXiv Detail & Related papers (2022-07-12T15:43:21Z) - Florida Wildlife Camera Trap Dataset [48.99466876948454]
We introduce a challenging wildlife camera trap classification dataset collected from two different locations in Southwestern Florida.
The dataset consists of 104,495 images featuring visually similar species, varying illumination conditions, skewed class distribution, and including samples of endangered species.
arXiv Detail & Related papers (2021-06-23T18:53:15Z) - AcinoSet: A 3D Pose Estimation Dataset and Baseline Models for Cheetahs
in the Wild [51.35013619649463]
We present an extensive dataset of free-running cheetahs in the wild, called AcinoSet.
The dataset contains 119,490 frames of multi-view synchronized high-speed video footage, camera calibration files and 7,588 human-annotated frames.
The resulting 3D trajectories, human-checked 3D ground truth, and an interactive tool to inspect the data is also provided.
arXiv Detail & Related papers (2021-03-24T15:54:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.