AP-10K: A Benchmark for Animal Pose Estimation in the Wild
- URL: http://arxiv.org/abs/2108.12617v1
- Date: Sat, 28 Aug 2021 10:23:34 GMT
- Title: AP-10K: A Benchmark for Animal Pose Estimation in the Wild
- Authors: Hang Yu, Yufei Xu, Jing Zhang, Wei Zhao, Ziyu Guan, Dacheng Tao
- Abstract summary: We propose AP-10K, the first large-scale benchmark for general animal pose estimation.
AP-10K consists of 10,015 images collected and filtered from 23 animal families and 60 species.
Results provide sound empirical evidence on the superiority of learning from diverse animals species in terms of both accuracy and generalization ability.
- Score: 83.17759850662826
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Accurate animal pose estimation is an essential step towards understanding
animal behavior, and can potentially benefit many downstream applications, such
as wildlife conservation. Previous works only focus on specific animals while
ignoring the diversity of animal species, limiting the generalization ability.
In this paper, we propose AP-10K, the first large-scale benchmark for general
animal pose estimation, to facilitate the research in animal pose estimation.
AP-10K consists of 10,015 images collected and filtered from 23 animal families
and 60 species following the taxonomic rank and high-quality keypoint
annotations labeled and checked manually. Based on AP-10K, we benchmark
representative pose estimation models on the following three tracks: (1)
supervised learning for animal pose estimation, (2) cross-domain transfer
learning from human pose estimation to animal pose estimation, and (3) intra-
and inter-family domain generalization for unseen animals. The experimental
results provide sound empirical evidence on the superiority of learning from
diverse animals species in terms of both accuracy and generalization ability.
It opens new directions for facilitating future research in animal pose
estimation. AP-10k is publicly available at
https://github.com/AlexTheBad/AP10K.
Related papers
- PoseBench: Benchmarking the Robustness of Pose Estimation Models under Corruptions [57.871692507044344]
Pose estimation aims to accurately identify anatomical keypoints in humans and animals using monocular images.
Current models are typically trained and tested on clean data, potentially overlooking the corruption during real-world deployment.
We introduce PoseBench, a benchmark designed to evaluate the robustness of pose estimation models against real-world corruption.
arXiv Detail & Related papers (2024-06-20T14:40:17Z) - APTv2: Benchmarking Animal Pose Estimation and Tracking with a
Large-scale Dataset and Beyond [27.50166679588048]
APTv2 is the pioneering large-scale benchmark for animal pose estimation and tracking.
It comprises 2,749 video clips filtered and collected from 30 distinct animal species.
We provide high-quality keypoint and tracking annotations for a total of 84,611 animal instances.
arXiv Detail & Related papers (2023-12-25T04:49:49Z) - MammalNet: A Large-scale Video Benchmark for Mammal Recognition and
Behavior Understanding [38.3767550066302]
MammalNet is a large-scale animal behavior dataset with taxonomy-guided annotations of mammals and their common behaviors.
It contains over 18K videos totaling 539 hours, which is 10 times larger than the largest existing animal behavior dataset.
We establish three benchmarks on MammalNet: standard animal and behavior recognition, compositional low-shot animal and behavior recognition, and behavior detection.
arXiv Detail & Related papers (2023-06-01T11:45:33Z) - CLAMP: Prompt-based Contrastive Learning for Connecting Language and
Animal Pose [70.59906971581192]
We introduce a novel prompt-based Contrastive learning scheme for connecting Language and AniMal Pose effectively.
The CLAMP attempts to bridge the gap by adapting the text prompts to the animal keypoints during network training.
Experimental results show that our method achieves state-of-the-art performance under the supervised, few-shot, and zero-shot settings.
arXiv Detail & Related papers (2022-06-23T14:51:42Z) - APT-36K: A Large-scale Benchmark for Animal Pose Estimation and Tracking [77.87449881852062]
APT-36K is the first large-scale benchmark for animal pose estimation and tracking.
It consists of 2,400 video clips collected and filtered from 30 animal species with 15 frames for each video, resulting in 36,000 frames in total.
We benchmark several representative models on the following three tracks: (1) supervised animal pose estimation on a single frame under intra- and inter-domain transfer learning settings, (2) inter-species domain generalization test for unseen animals, and (3) animal pose estimation with animal tracking.
arXiv Detail & Related papers (2022-06-12T07:18:36Z) - Animal Kingdom: A Large and Diverse Dataset for Animal Behavior
Understanding [4.606145900630665]
We create a large and diverse dataset, Animal Kingdom, that provides multiple annotated tasks.
Our dataset contains 50 hours of annotated videos to localize relevant animal behavior segments.
We propose a Collaborative Action Recognition (CARe) model that learns general and specific features for action recognition with unseen new animals.
arXiv Detail & Related papers (2022-04-18T02:05:15Z) - A Unified Framework for Domain Adaptive Pose Estimation [70.54942818742]
We propose a unified framework that generalizes well on various domain adaptive pose estimation problems.
Our method outperforms existing baselines on human pose estimation by up to 4.5 percent points (pp), hand pose estimation by up to 7.4 pp, and animal pose estimation by up to 4.8 pp for dogs and 3.3 pp for sheep.
arXiv Detail & Related papers (2022-04-01T02:47:31Z) - Pose Recognition in the Wild: Animal pose estimation using Agglomerative
Clustering and Contrastive Learning [0.0]
We introduce a novel architecture that is able to recognize the pose of multiple animals fromunlabelled data.
We are able to distinguish between body parts of the animal, based on their visual behavior, instead of the underlying anatomy.
arXiv Detail & Related papers (2021-11-16T07:00:31Z) - Transferring Dense Pose to Proximal Animal Classes [83.84439508978126]
We show that it is possible to transfer the knowledge existing in dense pose recognition for humans, as well as in more general object detectors and segmenters, to the problem of dense pose recognition in other classes.
We do this by establishing a DensePose model for the new animal which is also geometrically aligned to humans.
We also introduce two benchmark datasets labelled in the manner of DensePose for the class chimpanzee and use them to evaluate our approach.
arXiv Detail & Related papers (2020-02-28T21:43:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.