Pose Recognition in the Wild: Animal pose estimation using Agglomerative
Clustering and Contrastive Learning
- URL: http://arxiv.org/abs/2111.08259v1
- Date: Tue, 16 Nov 2021 07:00:31 GMT
- Title: Pose Recognition in the Wild: Animal pose estimation using Agglomerative
Clustering and Contrastive Learning
- Authors: Samayan Bhattacharya, Sk Shahnawaz
- Abstract summary: We introduce a novel architecture that is able to recognize the pose of multiple animals fromunlabelled data.
We are able to distinguish between body parts of the animal, based on their visual behavior, instead of the underlying anatomy.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Animal pose estimation has recently come into the limelight due to its
application in biology, zoology, and aquaculture. Deep learning methods have
effectively been applied to human pose estimation. However, the major
bottleneck to the application of these methods to animal pose estimation is the
unavailability of sufficient quantities of labeled data. Though there are ample
quantities of unlabelled data publicly available, it is economically
impractical to label large quantities of data for each animal. In addition, due
to the wide variety of body shapes in the animal kingdom, the transfer of
knowledge across domains is ineffective. Given the fact that the human brain is
able to recognize animal pose without requiring large amounts of labeled data,
it is only reasonable that we exploit unsupervised learning to tackle the
problem of animal pose recognition from the available, unlabelled data. In this
paper, we introduce a novel architecture that is able to recognize the pose of
multiple animals fromunlabelled data. We do this by (1) removing background
information from each image and employing an edge detection algorithm on the
body of the animal, (2) Tracking motion of the edge pixels and performing
agglomerative clustering to segment body parts, (3) employing contrastive
learning to discourage grouping of distant body parts together. Hence we are
able to distinguish between body parts of the animal, based on their visual
behavior, instead of the underlying anatomy. Thus, we are able to achieve a
more effective classification of the data than their human-labeled
counterparts. We test our model on the TigDog and WLD (WildLife Documentary)
datasets, where we outperform state-of-the-art approaches by a significant
margin. We also study the performance of our model on other public data to
demonstrate the generalization ability of our model.
Related papers
- PoseBench: Benchmarking the Robustness of Pose Estimation Models under Corruptions [57.871692507044344]
Pose estimation aims to accurately identify anatomical keypoints in humans and animals using monocular images.
Current models are typically trained and tested on clean data, potentially overlooking the corruption during real-world deployment.
We introduce PoseBench, a benchmark designed to evaluate the robustness of pose estimation models against real-world corruption.
arXiv Detail & Related papers (2024-06-20T14:40:17Z) - Multimodal Foundation Models for Zero-shot Animal Species Recognition in
Camera Trap Images [57.96659470133514]
Motion-activated camera traps constitute an efficient tool for tracking and monitoring wildlife populations across the globe.
Supervised learning techniques have been successfully deployed to analyze such imagery, however training such techniques requires annotations from experts.
Reducing the reliance on costly labelled data has immense potential in developing large-scale wildlife tracking solutions with markedly less human labor.
arXiv Detail & Related papers (2023-11-02T08:32:00Z) - Of Mice and Pose: 2D Mouse Pose Estimation from Unlabelled Data and
Synthetic Prior [0.7499722271664145]
We propose an approach for estimating 2D mouse body pose from unlabelled images using a synthetically generated empirical pose prior.
We adapt this method to the limb structure of the mouse and generate the empirical prior of 2D poses from a synthetic 3D mouse model.
In experiments on a new mouse video dataset, we evaluate the performance of the approach by comparing pose predictions to a manually obtained ground truth.
arXiv Detail & Related papers (2023-07-25T09:31:55Z) - Mine yOur owN Anatomy: Revisiting Medical Image Segmentation with Extremely Limited Labels [54.58539616385138]
We introduce a novel semi-supervised 2D medical image segmentation framework termed Mine yOur owN Anatomy (MONA)
First, prior work argues that every pixel equally matters to the model training; we observe empirically that this alone is unlikely to define meaningful anatomical features.
Second, we construct a set of objectives that encourage the model to be capable of decomposing medical images into a collection of anatomical features.
arXiv Detail & Related papers (2022-09-27T15:50:31Z) - Prior-Aware Synthetic Data to the Rescue: Animal Pose Estimation with
Very Limited Real Data [18.06492246414256]
We present a data efficient strategy for pose estimation in quadrupeds that requires only a small amount of real images from the target animal.
It is confirmed that fine-tuning a backbone network with pretrained weights on generic image datasets such as ImageNet can mitigate the high demand for target animal pose data.
We introduce a prior-aware synthetic animal data generation pipeline called PASyn to augment the animal pose data essential for robust pose estimation.
arXiv Detail & Related papers (2022-08-30T01:17:50Z) - SyDog: A Synthetic Dog Dataset for Improved 2D Pose Estimation [3.411873646414169]
SyDog is a synthetic dataset of dogs containing ground truth pose and bounding box coordinates.
We demonstrate that pose estimation models trained on SyDog achieve better performance than models trained purely on real data.
arXiv Detail & Related papers (2021-07-31T14:34:40Z) - Unsupervised Shape and Pose Disentanglement for 3D Meshes [49.431680543840706]
We present a simple yet effective approach to learn disentangled shape and pose representations in an unsupervised setting.
We use a combination of self-consistency and cross-consistency constraints to learn pose and shape space from registered meshes.
We demonstrate the usefulness of learned representations through a number of tasks including pose transfer and shape retrieval.
arXiv Detail & Related papers (2020-07-22T11:00:27Z) - Transferring Dense Pose to Proximal Animal Classes [83.84439508978126]
We show that it is possible to transfer the knowledge existing in dense pose recognition for humans, as well as in more general object detectors and segmenters, to the problem of dense pose recognition in other classes.
We do this by establishing a DensePose model for the new animal which is also geometrically aligned to humans.
We also introduce two benchmark datasets labelled in the manner of DensePose for the class chimpanzee and use them to evaluate our approach.
arXiv Detail & Related papers (2020-02-28T21:43:53Z) - Deformation-aware Unpaired Image Translation for Pose Estimation on
Laboratory Animals [56.65062746564091]
We aim to capture the pose of neuroscience model organisms, without using any manual supervision, to study how neural circuits orchestrate behaviour.
Our key contribution is the explicit and independent modeling of appearance, shape and poses in an unpaired image translation framework.
We demonstrate improved pose estimation accuracy on Drosophila melanogaster (fruit fly), Caenorhabditis elegans (worm) and Danio rerio (zebrafish)
arXiv Detail & Related papers (2020-01-23T15:34:11Z) - Learning landmark guided embeddings for animal re-identification [15.356786390476591]
We propose to improve embedding learning by exploiting body landmarks information explicitly.
Body landmarks are provided to the input of a CNN as confidence heatmaps.
We evaluate the proposed method on a large synthetic dataset and a small real dataset.
arXiv Detail & Related papers (2020-01-09T01:31:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.