CzechLynx: A Dataset for Individual Identification and Pose Estimation of the Eurasian Lynx
- URL: http://arxiv.org/abs/2506.04931v1
- Date: Thu, 05 Jun 2025 12:05:43 GMT
- Title: CzechLynx: A Dataset for Individual Identification and Pose Estimation of the Eurasian Lynx
- Authors: Lukas Picek, Elisa Belotti, Michal Bojda, Ludek Bufka, Vojtech Cermak, Martin Dula, Rostislav Dvorak, Luboslav Hrdy, Miroslav Jirik, Vaclav Kocourek, Josefa Krausova, Jirı Labuda, Jakub Straka, Ludek Toman, Vlado Trulık, Martin Vana, Miroslav Kutal,
- Abstract summary: We introduce CzechLynx, the first large-scale, open-access dataset for individual identification, 2D pose estimation, and instance segmentation of the Eurasian lynx (Lynx lynx)<n>CzechLynx includes more than 30k camera trap images annotated with segmentation masks, identity labels, and 20-point skeletons and covers 219 unique individuals across 15 years of systematic monitoring in two geographically distinct regions: Southwest Bohemia and the Western Carpathians.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: We introduce CzechLynx, the first large-scale, open-access dataset for individual identification, 2D pose estimation, and instance segmentation of the Eurasian lynx (Lynx lynx). CzechLynx includes more than 30k camera trap images annotated with segmentation masks, identity labels, and 20-point skeletons and covers 219 unique individuals across 15 years of systematic monitoring in two geographically distinct regions: Southwest Bohemia and the Western Carpathians. To increase the data variability, we create a complementary synthetic set with more than 100k photorealistic images generated via a Unity-based pipeline and diffusion-driven text-to-texture modeling, covering diverse environments, poses, and coat-pattern variations. To allow testing generalization across spatial and temporal domains, we define three tailored evaluation protocols/splits: (i) geo-aware, (ii) time-aware open-set, and (iii) time-aware closed-set. This dataset is targeted to be instrumental in benchmarking state-of-the-art models and the development of novel methods for not just individual animal re-identification.
Related papers
- A Dataset for Semantic Segmentation in the Presence of Unknowns [49.795683850385956]
Existing datasets allow evaluation of only knowns or unknowns - but not both.<n>We propose a novel anomaly segmentation dataset, ISSU, that features a diverse set of anomaly inputs from cluttered real-world environments.<n>The dataset is twice larger than existing anomaly segmentation datasets.
arXiv Detail & Related papers (2025-03-28T10:31:01Z) - Multispecies Animal Re-ID Using a Large Community-Curated Dataset [0.19418036471925312]
We construct a dataset that includes 49 species, 37K individual animals, and 225K images, using this data to train a single embedding network for all species.<n>Our model consistently outperforms models trained separately on each species, achieving an average gain of 12.5% in top-1 accuracy.<n>The model is already in production use for 60+ species in a large-scale wildlife monitoring system.
arXiv Detail & Related papers (2024-12-07T09:56:33Z) - CNN Based Flank Predictor for Quadruped Animal Species [1.502956022927019]
We train a flank predictor that predicts the visible flank of quadruped mammalian species in images.
The developed models were evaluated in different scenarios of different unknown quadruped species in known and unknown environments.
The best model, trained on an EfficientNetV2 backbone, achieved an accuracy of 88.70 % for the unknown species lynx in a complex habitat.
arXiv Detail & Related papers (2024-06-19T14:24:26Z) - SeaTurtleID2022: A long-span dataset for reliable sea turtle re-identification [0.0]
This paper introduces the first large-scale, long-span dataset with sea turtle photographs captured in the wild -- SeaTurtleID2022.
The dataset contains 8729 photographs of 438 unique individuals collected within 13 years.
Instead of standard "random" splits, the dataset allows for two realistic and ecologically motivated splits.
arXiv Detail & Related papers (2023-11-09T17:10:20Z) - IIITD-20K: Dense captioning for Text-Image ReID [5.858839403963778]
IIITD-20K comprises of 20,000 unique identities captured in the wild.
With a minimum of 26 words for a description, each image is densely captioned.
We perform elaborate experiments using state-of-art text-to-image ReID models and vision-language pre-trained models.
arXiv Detail & Related papers (2023-05-08T06:46:56Z) - Navya3DSeg -- Navya 3D Semantic Segmentation Dataset & split generation
for autonomous vehicles [63.20765930558542]
3D semantic data are useful for core perception tasks such as obstacle detection and ego-vehicle localization.
We propose a new dataset, Navya 3D (Navya3DSeg), with a diverse label space corresponding to a large scale production grade operational domain.
It contains 23 labeled sequences and 25 supplementary sequences without labels, designed to explore self-supervised and semi-supervised semantic segmentation benchmarks on point clouds.
arXiv Detail & Related papers (2023-02-16T13:41:19Z) - SeaTurtleID2022: A long-span dataset for reliable sea turtle re-identification [0.0]
This paper introduces the first public large-scale, long-span dataset with sea turtle photographs captured in the wild.
The dataset contains 8729 photographs of 438 unique individuals collected within 13 years.
arXiv Detail & Related papers (2022-11-18T15:46:24Z) - MSeg: A Composite Dataset for Multi-domain Semantic Segmentation [100.17755160696939]
We present MSeg, a composite dataset that unifies semantic segmentation datasets from different domains.
We reconcile the generalization and bring the pixel-level annotations into alignment by relabeling more than 220,000 object masks in more than 80,000 images.
A model trained on MSeg ranks first on the WildDash-v1 leaderboard for robust semantic segmentation, with no exposure to WildDash data during training.
arXiv Detail & Related papers (2021-12-27T16:16:35Z) - VoxelTrack: Multi-Person 3D Human Pose Estimation and Tracking in the
Wild [98.69191256693703]
We present VoxelTrack for multi-person 3D pose estimation and tracking from a few cameras which are separated by wide baselines.
It employs a multi-branch network to jointly estimate 3D poses and re-identification (Re-ID) features for all people in the environment.
It outperforms the state-of-the-art methods by a large margin on three public datasets including Shelf, Campus and CMU Panoptic.
arXiv Detail & Related papers (2021-08-05T08:35:44Z) - Shelf-Supervised Mesh Prediction in the Wild [54.01373263260449]
We propose a learning-based approach to infer 3D shape and pose of object from a single image.
We first infer a volumetric representation in a canonical frame, along with the camera pose.
The coarse volumetric prediction is then converted to a mesh-based representation, which is further refined in the predicted camera frame.
arXiv Detail & Related papers (2021-02-11T18:57:10Z) - Surpassing Real-World Source Training Data: Random 3D Characters for
Generalizable Person Re-Identification [109.68210001788506]
We propose to automatically synthesize a large-scale person re-identification dataset following a set-up similar to real surveillance.
We simulate a number of different virtual environments using Unity3D, with customized camera networks similar to real surveillance systems.
As a result, we obtain a virtual dataset, called RandPerson, with 1,801,816 person images of 8,000 identities.
arXiv Detail & Related papers (2020-06-23T05:38:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.