Identifying Individual Dogs in Social Media Images
- URL: http://arxiv.org/abs/2003.06705v1
- Date: Sat, 14 Mar 2020 21:11:02 GMT
- Title: Identifying Individual Dogs in Social Media Images
- Authors: Djordje Batic, Dubravko Culibrk
- Abstract summary: The work described here is part of joint project done with Pet2Net, a social network focused on pets and their owners.
In order to detect and recognize individual dogs we combine transfer learning and object detection approaches on Inception v3 and SSD Inception v2 architectures.
We show that it can achieve 94.59% accuracy in identifying individual dogs.
- Score: 1.14219428942199
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present the results of an initial study focused on developing a visual AI
solution able to recognize individual dogs in unconstrained (wild) images
occurring on social media.
The work described here is part of joint project done with Pet2Net, a social
network focused on pets and their owners. In order to detect and recognize
individual dogs we combine transfer learning and object detection approaches on
Inception v3 and SSD Inception v2 architectures respectively and evaluate the
proposed pipeline using a new data set containing real data that the users
uploaded to Pet2Net platform. We show that it can achieve 94.59% accuracy in
identifying individual dogs. Our approach has been designed with simplicity in
mind and the goal of easy deployment on all the images uploaded to Pet2Net
platform.
A purely visual approach to identifying dogs in images, will enhance Pet2Net
features aimed at finding lost dogs, as well as form the basis of future work
focused on identifying social relationships between dogs, which cannot be
inferred from other data collected by the platform.
Related papers
- The Dog Walking Theory: Rethinking Convergence in Federated Learning [68.63601691672815]
Federated learning (FL) is a collaborative learning paradigm that allows different clients to train one powerful global model without sharing their private data.
Although FL has demonstrated promising results in various applications, it is known to suffer from convergence issues caused by the data distribution shift across different clients.
We propose a novel FL algorithm emphFedWalk that leverages an external easy-to-converge task at the server side as a emphleash task to guide the local training of the clients.
arXiv Detail & Related papers (2024-04-18T04:25:21Z) - SocialPET: Socially Informed Pattern Exploiting Training for Few-Shot
Stance Detection in Social Media [8.556183465416156]
Stance detection is the task of determining the viewpoint of a social media post towards a target as 'favor' or 'against'
SocialPET is a socially informed approach to leveraging language models for the task.
We prove the effectiveness of SocialPET on two stance datasets, Multi-target and P-Stance.
arXiv Detail & Related papers (2024-03-08T11:00:09Z) - Categorizing the Visual Environment and Analyzing the Visual Attention
of Dogs [11.511035466613109]
We collect and study a dataset with over 11,698 gazes to categorize the objects available to be gazed at by 11 dogs.
A small portion of the collected data is used to fine tune a MaskRCNN for the novel image domain.
There are few individual differences between the 11 dogs and we observe greater visual fixations on buses, plants, pavement, and construction equipment.
arXiv Detail & Related papers (2023-11-20T18:21:18Z) - LostPaw: Finding Lost Pets using a Contrastive Learning-based
Transformer with Visual Input [0.5156484100374059]
This study introduces a contrastive neural network model capable of accurately distinguishing between images of pets.
The model was trained on a large dataset of dog images and evaluated through 3-fold cross-validation.
Our findings suggest that contrastive neural network models hold promise as a tool for locating lost pets.
arXiv Detail & Related papers (2023-04-28T11:23:44Z) - Video-based estimation of pain indicators in dogs [2.7103996289794217]
We propose a novel video-based, two-stream deep neural network approach for this problem.
We extract and preprocess body keypoints, and compute features from both keypoints and the RGB representation over the video.
We present a unique video-based dog behavior dataset, collected by veterinary professionals, and annotated for presence of pain.
arXiv Detail & Related papers (2022-09-27T10:38:59Z) - CLAMP: Prompt-based Contrastive Learning for Connecting Language and
Animal Pose [70.59906971581192]
We introduce a novel prompt-based Contrastive learning scheme for connecting Language and AniMal Pose effectively.
The CLAMP attempts to bridge the gap by adapting the text prompts to the animal keypoints during network training.
Experimental results show that our method achieves state-of-the-art performance under the supervised, few-shot, and zero-shot settings.
arXiv Detail & Related papers (2022-06-23T14:51:42Z) - A Competitive Method for Dog Nose-print Re-identification [46.94755073943372]
This paper presents our proposed methods for dog nose-print authentication (Re-ID) task in CVPR 2022 pet biometric challenge.
With multiple models ensembled adopted, our methods achieve 86.67% AUC on the test set.
arXiv Detail & Related papers (2022-05-31T16:26:46Z) - Persistent Animal Identification Leveraging Non-Visual Markers [71.14999745312626]
We aim to locate and provide a unique identifier for each mouse in a cluttered home-cage environment through time.
This is a very challenging problem due to (i) the lack of distinguishing visual features for each mouse, and (ii) the close confines of the scene with constant occlusion.
Our approach achieves 77% accuracy on this animal identification problem, and is able to reject spurious detections when the animals are hidden.
arXiv Detail & Related papers (2021-12-13T17:11:32Z) - Learning Co-segmentation by Segment Swapping for Retrieval and Discovery [67.6609943904996]
The goal of this work is to efficiently identify visually similar patterns from a pair of images.
We generate synthetic training pairs by selecting object segments in an image and copy-pasting them into another image.
We show our approach provides clear improvements for artwork details retrieval on the Brueghel dataset.
arXiv Detail & Related papers (2021-10-29T16:51:16Z) - Onfocus Detection: Identifying Individual-Camera Eye Contact from
Unconstrained Images [81.64699115587167]
Onfocus detection aims at identifying whether the focus of the individual captured by a camera is on the camera or not.
We build a large-scale onfocus detection dataset, named as the OnFocus Detection In the Wild (OFDIW)
We propose a novel end-to-end deep model, i.e., the eye-context interaction inferring network (ECIIN) for onfocus detection.
arXiv Detail & Related papers (2021-03-29T03:29:09Z) - Dog Identification using Soft Biometrics and Neural Networks [1.2922946578413577]
We apply advanced machine learning models such as deep neural network on the photographs of pets in order to determine the pet identity.
We explore the possibility of using different types of "soft" biometrics, such as breed, height, or gender, in fusion with "hard" biometrics such as photographs of the pet's face.
The proposed network is able to achieve an accuracy of 90.80% and 91.29% when differentiating between the two dog breeds.
arXiv Detail & Related papers (2020-07-22T10:22:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.