MMLA: Multi-Environment, Multi-Species, Low-Altitude Aerial Footage Dataset
- URL: http://arxiv.org/abs/2504.07744v1
- Date: Thu, 10 Apr 2025 13:40:27 GMT
- Title: MMLA: Multi-Environment, Multi-Species, Low-Altitude Aerial Footage Dataset
- Authors: Jenna Kline, Samuel Stevens, Guy Maalouf, Camille Rondeau Saint-Jean, Dat Nguyen Ngoc, Majid Mirmehdi, David Guerin, Tilo Burghardt, Elzbieta Pastucha, Blair Costelloe, Matthew Watson, Thomas Richardson, Ulrik Pagh Schultz Lundquist,
- Abstract summary: Real-time wildlife detection in drone imagery is critical for numerous applications, including animal ecology, conservation, and biodiversity monitoring.<n>We present a novel multi-species, multi-environment, low-altitude aerial footage (MMLA) dataset.<n>Results demonstrate significant performance disparities across locations and species-specific detection variations.
- Score: 3.7188931723069443
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Real-time wildlife detection in drone imagery is critical for numerous applications, including animal ecology, conservation, and biodiversity monitoring. Low-altitude drone missions are effective for collecting fine-grained animal movement and behavior data, particularly if missions are automated for increased speed and consistency. However, little work exists on evaluating computer vision models on low-altitude aerial imagery and generalizability across different species and settings. To fill this gap, we present a novel multi-environment, multi-species, low-altitude aerial footage (MMLA) dataset. MMLA consists of drone footage collected across three diverse environments: Ol Pejeta Conservancy and Mpala Research Centre in Kenya, and The Wilds Conservation Center in Ohio, which includes five species: Plains zebras, Grevy's zebras, giraffes, onagers, and African Painted Dogs. We comprehensively evaluate three YOLO models (YOLOv5m, YOLOv8m, and YOLOv11m) for detecting animals. Results demonstrate significant performance disparities across locations and species-specific detection variations. Our work highlights the importance of evaluating detection algorithms across different environments for robust wildlife monitoring applications using drones.
Related papers
- MVA 2025 Small Multi-Object Tracking for Spotting Birds Challenge: Dataset, Methods, and Results [15.90859212645041]
This paper introduces the SMOT4SB challenge, which leverages temporal information to address limitations of single-frame detection.<n>Our three main contributions are: (1) the SMOT4SB dataset, consisting of 211 UAV video sequences with 108,192 annotated frames under diverse real-world conditions; (2) SO-HOTA, a novel metric combining Dot Distance with HOTA to mitigate the sensitivity of IoU-based metrics to small displacements; and (3) a competitive MVA2025 challenge with 78 participants and 308 submissions, where the winning method achieved a 5.1x improvement over the baseline.
arXiv Detail & Related papers (2025-07-17T06:45:47Z) - WildLive: Near Real-time Visual Wildlife Tracking onboard UAVs [4.215854427679142]
We introduce WildLive -- a near real-time animal detection and tracking framework for high-resolution imagery running directly onboard aerial vehicles (UAVs)
The system performs multi-animal detection tracking at 17 fps+ for HD and 7 fps+ on 4K video streams suitable for operation during higher altitude flights to minimise animal disturbance.
We introduce our WildLive dataset, which comprises 200k+ annotated animal instances across 19k+ frames from 4K UAV videos collected at the Ol Pejeta Conservancy in Kenya.
arXiv Detail & Related papers (2025-04-14T12:21:16Z) - MammAlps: A multi-view video behavior monitoring dataset of wild mammals in the Swiss Alps [41.58000025132071]
MammAlps is a dataset of wildlife behavior monitoring from 9 camera-traps in the Swiss National Park.<n>Based on 6135 single animal clips, we propose the first hierarchical and multimodal animal behavior recognition benchmark.<n>We also propose a second ecology-oriented benchmark aiming at identifying activities, species, number of individuals and meteorological conditions.
arXiv Detail & Related papers (2025-03-23T21:51:58Z) - YOLOMG: Vision-based Drone-to-Drone Detection with Appearance and Pixel-Level Motion Fusion [9.810747004677474]
This paper proposes a novel end-to-end framework that accurately identifies small drones in complex environments.<n>It starts by creating a motion difference map to capture the motion characteristics of tiny drones.<n>Next, this motion difference map is combined with an RGB image using a bimodal fusion module, allowing for adaptive feature learning of the drone.
arXiv Detail & Related papers (2025-03-10T09:44:21Z) - Few-shot Species Range Estimation [61.60698161072356]
Knowing where a particular species can or cannot be found on Earth is crucial for ecological research and conservation efforts.<n>We outline a new approach for few-shot species range estimation to address the challenge of accurately estimating the range of a species from limited data.<n>During inference, our model takes a set of spatial locations as input, along with optional metadata such as text or an image, and outputs a species encoding that can be used to predict the range of a previously unseen species in feed-forward manner.
arXiv Detail & Related papers (2025-02-20T19:13:29Z) - BuckTales : A multi-UAV dataset for multi-object tracking and re-identification of wild antelopes [0.6267336085190178]
BuckTales is the first large-scale UAV dataset designed to solve multi-object tracking and re-identification problem in wild animals.
The MOT dataset includes over 1.2 million annotations including 680 tracks across 12 high-resolution (5.4K) videos.
The Re-ID dataset includes 730 individuals captured with two UAVs simultaneously.
arXiv Detail & Related papers (2024-11-11T11:55:14Z) - Multi-Species Object Detection in Drone Imagery for Population Monitoring of Endangered Animals [0.0]
This research focused on fine-tuning object detection models for drone images to create accurate counts of animal species.
We trained 30 different models, with the largest having 43.7 million parameters and 365 layers.
We deployed the models on the Jetson Orin Nano for demonstration of low-power real-time species detection for easy inference on drones.
arXiv Detail & Related papers (2024-06-28T03:03:55Z) - Multiview Aerial Visual Recognition (MAVREC): Can Multi-view Improve
Aerial Visual Perception? [57.77643186237265]
We present Multiview Aerial Visual RECognition or MAVREC, a video dataset where we record synchronized scenes from different perspectives.
MAVREC consists of around 2.5 hours of industry-standard 2.7K resolution video sequences, more than 0.5 million frames, and 1.1 million annotated bounding boxes.
This makes MAVREC the largest ground and aerial-view dataset, and the fourth largest among all drone-based datasets.
arXiv Detail & Related papers (2023-12-07T18:59:14Z) - Multimodal Foundation Models for Zero-shot Animal Species Recognition in
Camera Trap Images [57.96659470133514]
Motion-activated camera traps constitute an efficient tool for tracking and monitoring wildlife populations across the globe.
Supervised learning techniques have been successfully deployed to analyze such imagery, however training such techniques requires annotations from experts.
Reducing the reliance on costly labelled data has immense potential in developing large-scale wildlife tracking solutions with markedly less human labor.
arXiv Detail & Related papers (2023-11-02T08:32:00Z) - SatBird: Bird Species Distribution Modeling with Remote Sensing and
Citizen Science Data [68.2366021016172]
We present SatBird, a satellite dataset of locations in the USA with labels derived from presence-absence observation data from the citizen science database eBird.
We also provide a dataset in Kenya representing low-data regimes.
We benchmark a set of baselines on our dataset, including SOTA models for remote sensing tasks.
arXiv Detail & Related papers (2023-11-02T02:00:27Z) - SOCRATES: A Stereo Camera Trap for Monitoring of Biodiversity [0.0]
This study presents a novel approach to 3D camera trapping featuring highly optimized hardware and software.
A comprehensive evaluation of SOCRATES shows not only a $3.23%$ improvement in animal detection but also its superior applicability for estimating animal abundance.
arXiv Detail & Related papers (2022-09-19T15:03:35Z) - LudVision -- Remote Detection of Exotic Invasive Aquatic Floral Species
using Drone-Mounted Multispectral Data [0.34998703934432673]
Ludwigia peploides is considered by the European Union as an aquatic invasive species.
Our goal was to develop a method to identify the presence of the species.
To identify the targeted species on the collected images, we propose a new method for detecting Ludwigia p.
arXiv Detail & Related papers (2022-07-12T15:43:21Z) - A dataset for multi-sensor drone detection [67.75999072448555]
The use of small and remotely controlled unmanned aerial vehicles (UAVs) has increased in recent years.
Most studies on drone detection fail to specify the type of acquisition device, the drone type, the detection range, or the dataset.
We contribute with an annotated multi-sensor database for drone detection that includes infrared and visible videos and audio files.
arXiv Detail & Related papers (2021-11-02T20:52:03Z) - Florida Wildlife Camera Trap Dataset [48.99466876948454]
We introduce a challenging wildlife camera trap classification dataset collected from two different locations in Southwestern Florida.
The dataset consists of 104,495 images featuring visually similar species, varying illumination conditions, skewed class distribution, and including samples of endangered species.
arXiv Detail & Related papers (2021-06-23T18:53:15Z) - Detection, Tracking, and Counting Meets Drones in Crowds: A Benchmark [97.07865343576361]
We construct a benchmark with a new drone-captured largescale dataset, named as DroneCrowd.
We annotate 20,800 people trajectories with 4.8 million heads and several video-level attributes.
We design the Space-Time Neighbor-Aware Network (STNNet) as a strong baseline to solve object detection, tracking and counting jointly in dense crowds.
arXiv Detail & Related papers (2021-05-06T04:46:14Z) - AcinoSet: A 3D Pose Estimation Dataset and Baseline Models for Cheetahs
in the Wild [51.35013619649463]
We present an extensive dataset of free-running cheetahs in the wild, called AcinoSet.
The dataset contains 119,490 frames of multi-view synchronized high-speed video footage, camera calibration files and 7,588 human-annotated frames.
The resulting 3D trajectories, human-checked 3D ground truth, and an interactive tool to inspect the data is also provided.
arXiv Detail & Related papers (2021-03-24T15:54:11Z) - Multi-Drone based Single Object Tracking with Agent Sharing Network [74.8198920355117]
Multi-Drone single Object Tracking dataset consists of 92 groups of video clips with 113,918 high resolution frames taken by two drones and 63 groups of video clips with 145,875 high resolution frames taken by three drones.
Agent sharing network (ASNet) is proposed by self-supervised template sharing and view-aware fusion of the target from multiple drones.
arXiv Detail & Related papers (2020-03-16T03:27:04Z) - Detection and Tracking Meet Drones Challenge [131.31749447313197]
This paper presents a review of object detection and tracking datasets and benchmarks, and discusses the challenges of collecting large-scale drone-based object detection and tracking datasets with manual annotations.
We describe our VisDrone dataset, which is captured over various urban/suburban areas of 14 different cities across China from North to South.
We provide a detailed analysis of the current state of the field of large-scale object detection and tracking on drones, and conclude the challenge as well as propose future directions.
arXiv Detail & Related papers (2020-01-16T00:11:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.