DogFLW: Dog Facial Landmarks in the Wild Dataset
- URL: http://arxiv.org/abs/2405.11501v1
- Date: Sun, 19 May 2024 09:59:36 GMT
- Title: DogFLW: Dog Facial Landmarks in the Wild Dataset
- Authors: George Martvel, Greta Abele, Annika Bremhorst, Chiara Canori, Nareed Farhat, Giulia Pedretti, Ilan Shimshoni, Anna Zamansky,
- Abstract summary: We develop an analogous dataset containing 3,274 annotated images of dogs.
Our dataset is based on a scheme of 46 facial anatomy-based landmarks.
- Score: 5.51724397504775
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Affective computing for animals is a rapidly expanding research area that is going deeper than automated movement tracking to address animal internal states, like pain and emotions. Facial expressions can serve to communicate information about these states in mammals. However, unlike human-related studies, there is a significant shortage of datasets that would enable the automated analysis of animal facial expressions. Inspired by the recently introduced Cat Facial Landmarks in the Wild dataset, presenting cat faces annotated with 48 facial anatomy-based landmarks, in this paper, we develop an analogous dataset containing 3,274 annotated images of dogs. Our dataset is based on a scheme of 46 facial anatomy-based landmarks. The DogFLW dataset is available from the corresponding author upon a reasonable request.
Related papers
- PetFace: A Large-Scale Dataset and Benchmark for Animal Identification [2.3020018305241337]
We introduce the PetFace dataset, a comprehensive resource for animal face identification.
PetFace includes 257,484 unique individuals across 13 animal families and 319 breed categories, including both experimental and pet animals.
We provide benchmarks including re-identification for seen individuals and verification for unseen individuals.
arXiv Detail & Related papers (2024-07-18T14:28:31Z) - OmniMotionGPT: Animal Motion Generation with Limited Data [70.35662376853163]
We introduce AnimalML3D, the first text-animal motion dataset with 1240 animation sequences spanning 36 different animal identities.
We are able to generate animal motions with high diversity and fidelity, quantitatively and qualitatively outperforming the results of training human motion generation baselines on animal data.
arXiv Detail & Related papers (2023-11-30T07:14:00Z) - Categorizing the Visual Environment and Analyzing the Visual Attention
of Dogs [11.511035466613109]
We collect and study a dataset with over 11,698 gazes to categorize the objects available to be gazed at by 11 dogs.
A small portion of the collected data is used to fine tune a MaskRCNN for the novel image domain.
There are few individual differences between the 11 dogs and we observe greater visual fixations on buses, plants, pavement, and construction equipment.
arXiv Detail & Related papers (2023-11-20T18:21:18Z) - Automated Detection of Cat Facial Landmarks [8.435125986009881]
We present a novel dataset of cat facial images annotated with bounding boxes and 48 facial landmarks grounded in cat facial anatomy.
We introduce a landmark detection convolution neural network-based model which uses a magnifying ensembe method.
Our model shows excellent performance on cat faces and is generalizable to human facial landmark detection.
arXiv Detail & Related papers (2023-10-15T10:44:36Z) - CatFLW: Cat Facial Landmarks in the Wild Dataset [3.3022383854506843]
This paper presents a dataset called Cat Facial Landmarks in the Wild (CatFLW)
It contains 2016 images of cat faces in different environments and conditions, annotated with 48 facial landmarks.
To the best of our knowledge, this dataset has the largest amount of cat facial landmarks available.
arXiv Detail & Related papers (2023-05-07T09:39:12Z) - Multi-Domain Norm-referenced Encoding Enables Data Efficient Transfer
Learning of Facial Expression Recognition [62.997667081978825]
We propose a biologically-inspired mechanism for transfer learning in facial expression recognition.
Our proposed architecture provides an explanation for how the human brain might innately recognize facial expressions on varying head shapes.
Our model achieves a classification accuracy of 92.15% on the FERG dataset with extreme data efficiency.
arXiv Detail & Related papers (2023-04-05T09:06:30Z) - Are Face Detection Models Biased? [69.68854430664399]
We investigate possible bias in the domain of face detection through facial region localization.
Most existing face detection datasets lack suitable annotation for such analysis.
We observe a high disparity in detection accuracies across gender and skin-tone, and interplay of confounding factors beyond demography.
arXiv Detail & Related papers (2022-11-07T14:27:55Z) - APT-36K: A Large-scale Benchmark for Animal Pose Estimation and Tracking [77.87449881852062]
APT-36K is the first large-scale benchmark for animal pose estimation and tracking.
It consists of 2,400 video clips collected and filtered from 30 animal species with 15 frames for each video, resulting in 36,000 frames in total.
We benchmark several representative models on the following three tracks: (1) supervised animal pose estimation on a single frame under intra- and inter-domain transfer learning settings, (2) inter-species domain generalization test for unseen animals, and (3) animal pose estimation with animal tracking.
arXiv Detail & Related papers (2022-06-12T07:18:36Z) - Deep Learning Models for Automated Classification of Dog Emotional
States from Facial Expressions [1.32383730641561]
We apply recent deep learning techniques to classify (positive) anticipation and (negative) frustration of dogs.
To the best of our knowledge, this work is the first to address the task of automatic classification of canine emotions.
arXiv Detail & Related papers (2022-06-11T21:37:38Z) - Overcoming the Domain Gap in Neural Action Representations [60.47807856873544]
3D pose data can now be reliably extracted from multi-view video sequences without manual intervention.
We propose to use it to guide the encoding of neural action representations together with a set of neural and behavioral augmentations.
To reduce the domain gap, during training, we swap neural and behavioral data across animals that seem to be performing similar actions.
arXiv Detail & Related papers (2021-12-02T12:45:46Z) - CapsField: Light Field-based Face and Expression Recognition in the Wild
using Capsule Routing [81.21490913108835]
This paper proposes a new deep face and expression recognition solution, called CapsField, based on a convolutional neural network.
The proposed solution achieves superior performance for both face and expression recognition tasks when compared to the state-of-the-art.
arXiv Detail & Related papers (2021-01-10T09:06:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.