BirdsEye-RU: A Dataset For Detecting Faces from Overhead Images
- URL: http://arxiv.org/abs/2601.12533v2
- Date: Wed, 21 Jan 2026 03:24:02 GMT
- Title: BirdsEye-RU: A Dataset For Detecting Faces from Overhead Images
- Authors: Md. Ahanaf Arif Khan, Ariful Islam, Sangeeta Biswas, Md. Iqbal Aziz Khan, Subrata Pramanik, Sanjoy Kumar Chakravarty, Bimal Kumar Pramanik,
- Abstract summary: The BirdsEye-RU dataset is a collection of 2,978 images containing over eight thousand annotated faces.<n>This dataset is specifically designed to capture small and distant faces across diverse environments.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Detecting faces in overhead images remains a significant challenge due to extreme scale variations and environmental clutter. To address this, we created the BirdsEye-RU dataset, a comprehensive collection of 2,978 images containing over eight thousand annotated faces. This dataset is specifically designed to capture small and distant faces across diverse environments, containing both drone images and smartphone-captured images from high altitude. We present a detailed description of the BirdsEye-RU dataset in this paper. We made our dataset freely available to the public, and it can be accessed at https://www.kaggle.com/datasets/mdahanafarifkhan/birdseye-ru.
Related papers
- Generalized Referring Expression Segmentation on Aerial Photos [47.944645462877894]
This work presents Aerial-D, a new large-scale referring expression segmentation dataset for aerial imagery.<n>It comprises 37,288 images with 1,522,523 referring expressions that cover 259,709 annotated targets, spanning across individual object instances, groups of instances, and semantic regions.<n>We adopted the RSRefSeg architecture, and trained models on Aerial-D together with prior aerial datasets, yielding unified instance and semantic segmentation from text for both modern and historical images.
arXiv Detail & Related papers (2025-12-08T09:25:59Z) - OLATverse: A Large-scale Real-world Object Dataset with Precise Lighting Control [65.57356151743936]
OLATverse is a large-scale dataset comprising around 9M images of 765 real-world objects.<n>Each object is captured using 35 DSLR cameras and 331 individually controlled light sources, enabling the simulation of diverse illumination conditions.<n>We construct an extensive evaluation set, establishing the first comprehensive real-world object-centric benchmark for inverse rendering and normal estimation.
arXiv Detail & Related papers (2025-11-04T11:13:43Z) - 360 in the Wild: Dataset for Depth Prediction and View Synthesis [66.58513725342125]
We introduce a large scale 360$circ$ videos dataset in the wild.
This dataset has been carefully scraped from the Internet and has been captured from various locations worldwide.
Each of the 25K images constituting our dataset is provided with its respective camera's pose and depth map.
arXiv Detail & Related papers (2024-06-27T05:26:38Z) - OpenIllumination: A Multi-Illumination Dataset for Inverse Rendering
Evaluation on Real Objects [56.065616159398324]
We introduce OpenIllumination, a real-world dataset containing over 108K images of 64 objects with diverse materials.
For each image in the dataset, we provide accurate camera parameters, illumination ground truth, and foreground segmentation masks.
arXiv Detail & Related papers (2023-09-14T17:59:53Z) - Ground-to-Aerial Person Search: Benchmark Dataset and Approach [42.54151390290665]
We construct a large-scale dataset for Ground-to-Aerial Person Search, named G2APS.
G2APS contains 31,770 images of 260,559 annotated bounding boxes for 2,644 identities appearing in both of the UAVs and ground surveillance cameras.
arXiv Detail & Related papers (2023-08-24T11:11:26Z) - VDD: Varied Drone Dataset for Semantic Segmentation [9.581655974280217]
We release a large-scale, densely labeled collection of 400 high-resolution images spanning 7 classes.<n>This dataset features various scenes in urban, industrial, rural, and natural areas, captured from different camera angles and under diverse lighting conditions.<n>We train seven state-of-the-art models on drone datasets as baselines.
arXiv Detail & Related papers (2023-05-23T02:16:14Z) - Booster: a Benchmark for Depth from Images of Specular and Transparent
Surfaces [49.44971010149331]
We propose a novel dataset that includes accurate and dense ground-truth labels at high resolution.
Our acquisition pipeline leverages a novel deep space-time stereo framework.
The dataset is composed of 606 samples collected in 85 different scenes.
arXiv Detail & Related papers (2023-01-19T18:59:28Z) - A Multi-purpose Real Haze Benchmark with Quantifiable Haze Levels and
Ground Truth [61.90504318229845]
This paper introduces the first paired real image benchmark dataset with hazy and haze-free images, and in-situ haze density measurements.
This dataset was produced in a controlled environment with professional smoke generating machines that covered the entire scene.
A subset of this dataset has been used for the Object Detection in Haze Track of CVPR UG2 2022 challenge.
arXiv Detail & Related papers (2022-06-13T19:14:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.