T-LEAP: occlusion-robust pose estimation of walking cows using temporal
information
- URL: http://arxiv.org/abs/2104.08029v1
- Date: Fri, 16 Apr 2021 10:50:56 GMT
- Title: T-LEAP: occlusion-robust pose estimation of walking cows using temporal
information
- Authors: Helena Russello, Rik van der Tol, Gert Kootstra
- Abstract summary: Lameness, a prevalent health disorder in dairy cows, is commonly detected by analyzing the gait of cows.
A cow's gait can be tracked in videos using pose estimation models because models learn to automatically localize anatomical landmarks in images and videos.
Most animal pose estimation models are static, that is, videos are processed frame by frame and do not use any temporal information.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As herd size on dairy farms continue to increase, automatic health monitoring
of cows has gained in interest. Lameness, a prevalent health disorder in dairy
cows, is commonly detected by analyzing the gait of cows. A cow's gait can be
tracked in videos using pose estimation models because models learn to
automatically localize anatomical landmarks in images and videos. Most animal
pose estimation models are static, that is, videos are processed frame by frame
and do not use any temporal information. In this work, a static deep-learning
model for animal-pose-estimation was extended to a temporal model that includes
information from past frames. We compared the performance of the static and
temporal pose estimation models. The data consisted of 1059 samples of 4
consecutive frames extracted from videos (30 fps) of 30 different dairy cows
walking through an outdoor passageway. As farm environments are prone to
occlusions, we tested the robustness of the static and temporal models by
adding artificial occlusions to the videos. The experiments showed that, on
non-occluded data, both static and temporal approaches achieved a Percentage of
Correct Keypoints (PCKh@0.2) of 99%. On occluded data, our temporal approach
outperformed the static one by up to 32.9%, suggesting that using temporal data
is beneficial for pose estimation in environments prone to occlusions, such as
dairy farms. The generalization capabilities of the temporal model was
evaluated by testing it on data containing unknown cows (cows not present in
the training set). The results showed that the average detection rate
(PCKh@0.2) was of 93.8% on known cows and 87.6% on unknown cows, indicating
that the model is capable of generalizing well to new cows and that they could
be easily fine-tuned to new herds. Finally, we showed that with harder tasks,
such as occlusions and unknown cows, a deeper architecture was more beneficial.
Related papers
- Supervised Learning Model for Key Frame Identification from Cow Teat Videos [0.9115927248875568]
This paper proposes a method for improving the accuracy of mastitis risk assessment in cows using neural networks and video analysis.
Traditionally, veterinarians assess the health of a cow's teat during the milking process.
This paper uses a neural network to identify key frames in the recorded video where the cow's udder appears intact.
arXiv Detail & Related papers (2024-09-26T15:50:43Z) - PoseBench: Benchmarking the Robustness of Pose Estimation Models under Corruptions [57.871692507044344]
Pose estimation aims to accurately identify anatomical keypoints in humans and animals using monocular images.
Current models are typically trained and tested on clean data, potentially overlooking the corruption during real-world deployment.
We introduce PoseBench, a benchmark designed to evaluate the robustness of pose estimation models against real-world corruption.
arXiv Detail & Related papers (2024-06-20T14:40:17Z) - CenTime: Event-Conditional Modelling of Censoring in Survival Analysis [49.44664144472712]
We introduce CenTime, a novel approach to survival analysis that directly estimates the time to event.
Our method features an innovative event-conditional censoring mechanism that performs robustly even when uncensored data is scarce.
Our results indicate that CenTime offers state-of-the-art performance in predicting time-to-death while maintaining comparable ranking performance.
arXiv Detail & Related papers (2023-09-07T17:07:33Z) - BovineTalk: Machine Learning for Vocalization Analysis of Dairy Cattle
under Negative Affective States [0.09786690381850353]
Cows were shown to produce two types vocalizations: low-frequency calls (LF), produced with the mouth closed, or partially closed, for close distance contacts, and high-frequency calls (HF), produced for long distance communication.
Here we present two computational frameworks - deep learning based and explainable machine learning based, to classify high and low-frequency cattle calls, and individual cow voice recognition.
arXiv Detail & Related papers (2023-07-26T07:07:03Z) - Depth video data-enabled predictions of longitudinal dairy cow body
weight using thresholding and Mask R-CNN algorithms [3.6809995018369936]
The aim of this study was to predict cow body weight from repeatedly measured video data.
Three approaches were investigated, including single thresholding, adaptive thresholding, and Mask R-CNN.
Our results suggest that deep learning-based segmentation improves the prediction performance of cow body weight from longitudinal depth video data.
arXiv Detail & Related papers (2023-07-03T22:27:37Z) - Particle-Based Score Estimation for State Space Model Learning in
Autonomous Driving [62.053071723903834]
Multi-object state estimation is a fundamental problem for robotic applications.
We consider learning maximum-likelihood parameters using particle methods.
We apply our method to real data collected from autonomous vehicles.
arXiv Detail & Related papers (2022-12-14T01:21:05Z) - Segmentation Enhanced Lameness Detection in Dairy Cows from RGB and
Depth Video [8.906235809404189]
Early lameness detection helps farmers address illnesses early and avoid negative effects caused by the degeneration of cows' condition.
We collected a dataset of short clips of cows exiting a milking station and annotated the degree of lameness of the cows.
We proposed a lameness detection method that leverages pre-trained neural networks to extract discriminative features from videos and assign a binary score to each cow indicating its condition: "healthy" or "lame"
arXiv Detail & Related papers (2022-06-09T12:16:31Z) - Using machine learning on new feature sets extracted from 3D models of
broken animal bones to classify fragments according to break agent [53.796331564067835]
We present a new approach to fracture pattern analysis aimed at distinguishing bone fragments resulting from hominin bone breakage and those produced by carnivores.
This new method uses 3D models of fragmentary bone to extract a much richer dataset that is more transparent and replicable than feature sets previously used in fracture pattern analysis.
Supervised machine learning algorithms are properly used to classify bone fragments according to agent of breakage with average mean accuracy of 77% across tests.
arXiv Detail & Related papers (2022-05-20T20:16:21Z) - Learning Summary Statistics for Bayesian Inference with Autoencoders [58.720142291102135]
We use the inner dimension of deep neural network based Autoencoders as summary statistics.
To create an incentive for the encoder to encode all the parameter-related information but not the noise, we give the decoder access to explicit or implicit information that has been used to generate the training data.
arXiv Detail & Related papers (2022-01-28T12:00:31Z) - Video-based cattle identification and action recognition [0.910162205512083]
We demonstrate a working prototype for the monitoring of cow welfare by automatically analysing the animal behaviours.
Deep learning models have been developed and tested with videos acquired in a farm, and a precision of 81.2% has been achieved for cow identification.
arXiv Detail & Related papers (2021-10-14T00:55:56Z) - Self-Attentive 3D Human Pose and Shape Estimation from Videos [82.63503361008607]
We present a video-based learning algorithm for 3D human pose and shape estimation.
We exploit temporal information in videos and propose a self-attention module.
We evaluate our method on the 3DPW, MPI-INF-3DHP, and Human3.6M datasets.
arXiv Detail & Related papers (2021-03-26T00:02:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.