The MONET dataset: Multimodal drone thermal dataset recorded in rural
scenarios
- URL: http://arxiv.org/abs/2304.05417v2
- Date: Wed, 19 Jul 2023 10:01:29 GMT
- Title: The MONET dataset: Multimodal drone thermal dataset recorded in rural
scenarios
- Authors: Luigi Riz, Andrea Caraffa, Matteo Bortolon, Mohamed Lamine Mekhalfi,
Davide Boscaini, Andr\'e Moura, Jos\'e Antunes, Andr\'e Dias, Hugo Silva,
Andreas Leonidou, Christos Constantinides, Christos Keleshis, Dante Abate,
Fabio Poiesi
- Abstract summary: We present MONET, a new multimodal dataset captured using a thermal camera mounted on a drone that flew over rural areas.
Monet consists of approximately 53K images featuring 162K manually annotated bounding boxes.
Each image is timestamp-aligned with drone metadata that includes information about attitudes, speed, altitude, and GPS coordinates.
- Score: 2.4683968227344097
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present MONET, a new multimodal dataset captured using a thermal camera
mounted on a drone that flew over rural areas, and recorded human and vehicle
activities. We captured MONET to study the problem of object localisation and
behaviour understanding of targets undergoing large-scale variations and being
recorded from different and moving viewpoints. Target activities occur in two
different land sites, each with unique scene structures and cluttered
backgrounds. MONET consists of approximately 53K images featuring 162K manually
annotated bounding boxes. Each image is timestamp-aligned with drone metadata
that includes information about attitudes, speed, altitude, and GPS
coordinates. MONET is different from previous thermal drone datasets because it
features multimodal data, including rural scenes captured with thermal cameras
containing both person and vehicle targets, along with trajectory information
and metadata. We assessed the difficulty of the dataset in terms of transfer
learning between the two sites and evaluated nine object detection algorithms
to identify the open challenges associated with this type of data. Project
page: https://github.com/fabiopoiesi/monet_dataset.
Related papers
- Multiview Aerial Visual Recognition (MAVREC): Can Multi-view Improve
Aerial Visual Perception? [57.77643186237265]
We present Multiview Aerial Visual RECognition or MAVREC, a video dataset where we record synchronized scenes from different perspectives.
MAVREC consists of around 2.5 hours of industry-standard 2.7K resolution video sequences, more than 0.5 million frames, and 1.1 million annotated bounding boxes.
This makes MAVREC the largest ground and aerial-view dataset, and the fourth largest among all drone-based datasets.
arXiv Detail & Related papers (2023-12-07T18:59:14Z) - VDD: Varied Drone Dataset for Semantic Segmentation [9.581655974280217]
We release a large-scale, densely labeled collection of 400 high-resolution images spanning 7 classes.
This dataset features various scenes in urban, industrial, rural, and natural areas, captured from different camera angles and under diverse lighting conditions.
We train seven state-of-the-art models on drone datasets as baselines.
arXiv Detail & Related papers (2023-05-23T02:16:14Z) - Synthehicle: Multi-Vehicle Multi-Camera Tracking in Virtual Cities [4.4855664250147465]
We present a massive synthetic dataset for multiple vehicle tracking and segmentation in multiple overlapping and non-overlapping camera views.
The dataset consists of 17 hours of labeled video material, recorded from 340 cameras in 64 diverse day, rain, dawn, and night scenes.
arXiv Detail & Related papers (2022-08-30T11:36:07Z) - EagerMOT: 3D Multi-Object Tracking via Sensor Fusion [68.8204255655161]
Multi-object tracking (MOT) enables mobile robots to perform well-informed motion planning and navigation by localizing surrounding objects in 3D space and time.
Existing methods rely on depth sensors (e.g., LiDAR) to detect and track targets in 3D space, but only up to a limited sensing range due to the sparsity of the signal.
We propose EagerMOT, a simple tracking formulation that integrates all available object observations from both sensor modalities to obtain a well-informed interpretation of the scene dynamics.
arXiv Detail & Related papers (2021-04-29T22:30:29Z) - A Flow Base Bi-path Network for Cross-scene Video Crowd Understanding in
Aerial View [93.23947591795897]
In this paper, we strive to tackle the challenges and automatically understand the crowd from the visual data collected from drones.
To alleviate the background noise generated in cross-scene testing, a double-stream crowd counting model is proposed.
To tackle the crowd density estimation problem under extreme dark environments, we introduce synthetic data generated by game Grand Theft Auto V(GTAV)
arXiv Detail & Related papers (2020-09-29T01:48:24Z) - Perceiving Traffic from Aerial Images [86.994032967469]
We propose an object detection method called Butterfly Detector that is tailored to detect objects in aerial images.
We evaluate our Butterfly Detector on two publicly available UAV datasets (UAVDT and VisDrone 2019) and show that it outperforms previous state-of-the-art methods while remaining real-time.
arXiv Detail & Related papers (2020-09-16T11:37:43Z) - Counting dense objects in remote sensing images [52.182698295053264]
Estimating number of interested objects from a given image is a challenging yet important task.
In this paper, we are interested in counting dense objects from remote sensing images.
To address these issues, we first construct a large-scale object counting dataset based on remote sensing images.
We then benchmark the dataset by designing a novel neural network which can generate density map of an input image.
arXiv Detail & Related papers (2020-02-14T09:13:54Z) - AU-AIR: A Multi-modal Unmanned Aerial Vehicle Dataset for Low Altitude
Traffic Surveillance [20.318367304051176]
Unmanned aerial vehicles (UAVs) with mounted cameras have the advantage of capturing aerial (bird-view) images.
Several aerial datasets have been introduced, including visual data with object annotations.
We propose a multi-purpose aerial dataset (AU-AIR) that has multi-modal sensor data collected in real-world outdoor environments.
arXiv Detail & Related papers (2020-01-31T09:45:12Z) - Detection and Tracking Meet Drones Challenge [131.31749447313197]
This paper presents a review of object detection and tracking datasets and benchmarks, and discusses the challenges of collecting large-scale drone-based object detection and tracking datasets with manual annotations.
We describe our VisDrone dataset, which is captured over various urban/suburban areas of 14 different cities across China from North to South.
We provide a detailed analysis of the current state of the field of large-scale object detection and tracking on drones, and conclude the challenge as well as propose future directions.
arXiv Detail & Related papers (2020-01-16T00:11:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.