BovineTalk: Machine Learning for Vocalization Analysis of Dairy Cattle
under Negative Affective States
- URL: http://arxiv.org/abs/2307.13994v1
- Date: Wed, 26 Jul 2023 07:07:03 GMT
- Title: BovineTalk: Machine Learning for Vocalization Analysis of Dairy Cattle
under Negative Affective States
- Authors: Dinu Gavojdian, Teddy Lazebnik, Madalina Mincu, Ariel Oren, Ioana
Nicolae, Anna Zamansky
- Abstract summary: Cows were shown to produce two types vocalizations: low-frequency calls (LF), produced with the mouth closed, or partially closed, for close distance contacts, and high-frequency calls (HF), produced for long distance communication.
Here we present two computational frameworks - deep learning based and explainable machine learning based, to classify high and low-frequency cattle calls, and individual cow voice recognition.
- Score: 0.09786690381850353
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: There is a critical need to develop and validate non-invasive animal-based
indicators of affective states in livestock species, in order to integrate them
into on-farm assessment protocols, potentially via the use of precision
livestock farming (PLF) tools. One such promising approach is the use of vocal
indicators. The acoustic structure of vocalizations and their functions were
extensively studied in important livestock species, such as pigs, horses,
poultry and goats, yet cattle remain understudied in this context to date. Cows
were shown to produce two types vocalizations: low-frequency calls (LF),
produced with the mouth closed, or partially closed, for close distance
contacts and open mouth emitted high-frequency calls (HF), produced for long
distance communication, with the latter considered to be largely associated
with negative affective states. Moreover, cattle vocalizations were shown to
contain information on individuality across a wide range of contexts, both
negative and positive. Nowadays, dairy cows are facing a series of negative
challenges and stressors in a typical production cycle, making vocalizations
during negative affective states of special interest for research. One
contribution of this study is providing the largest to date pre-processed
(clean from noises) dataset of lactating adult multiparous dairy cows during
negative affective states induced by visual isolation challenges. Here we
present two computational frameworks - deep learning based and explainable
machine learning based, to classify high and low-frequency cattle calls, and
individual cow voice recognition. Our models in these two frameworks reached
87.2% and 89.4% accuracy for LF and HF classification, with 68.9% and 72.5%
accuracy rates for the cow individual identification, respectively.
Related papers
- Multi Modal Information Fusion of Acoustic and Linguistic Data for Decoding Dairy Cow Vocalizations in Animal Welfare Assessment [0.0]
This study aims to decode dairy cow contact calls by employing multi-modal data fusion techniques.
We utilize the Natural Language Processing model to transcribe audio recordings of cow vocalizations into written form.
We categorized vocalizations into high frequency calls associated with distress or arousal, and low frequency calls linked to contentment or calmness.
arXiv Detail & Related papers (2024-11-01T09:48:30Z) - Supervised Learning Model for Key Frame Identification from Cow Teat Videos [0.9115927248875568]
This paper proposes a method for improving the accuracy of mastitis risk assessment in cows using neural networks and video analysis.
Traditionally, veterinarians assess the health of a cow's teat during the milking process.
This paper uses a neural network to identify key frames in the recorded video where the cow's udder appears intact.
arXiv Detail & Related papers (2024-09-26T15:50:43Z) - Measuring and Improving Attentiveness to Partial Inputs with Counterfactuals [91.59906995214209]
We propose a new evaluation method, Counterfactual Attentiveness Test (CAT)
CAT uses counterfactuals by replacing part of the input with its counterpart from a different example, expecting an attentive model to change its prediction.
We show that GPT3 becomes less attentive with an increased number of demonstrations, while its accuracy on the test data improves.
arXiv Detail & Related papers (2023-11-16T06:27:35Z) - Towards Automated Animal Density Estimation with Acoustic Spatial
Capture-Recapture [2.5193666094305938]
Digital recorders allow surveyors to gather large volumes of data at low cost.
But identifying target species vocalisations in these data is non-trivial.
Machine learning (ML) methods are often used to do the identification.
We propose three methods for acoustic spatial capture-recapture inference.
arXiv Detail & Related papers (2023-08-24T15:29:24Z) - Automatically measuring speech fluency in people with aphasia: first
achievements using read-speech data [55.84746218227712]
This study aims at assessing the relevance of a signalprocessingalgorithm, initially developed in the field of language acquisition, for the automatic measurement of speech fluency.
arXiv Detail & Related papers (2023-08-09T07:51:40Z) - Co-Learning Meets Stitch-Up for Noisy Multi-label Visual Recognition [70.00984078351927]
This paper focuses on reducing noise based on some inherent properties of multi-label classification and long-tailed learning under noisy cases.
We propose a Stitch-Up augmentation to synthesize a cleaner sample, which directly reduces multi-label noise.
A Heterogeneous Co-Learning framework is further designed to leverage the inconsistency between long-tailed and balanced distributions.
arXiv Detail & Related papers (2023-07-03T09:20:28Z) - Fairness Improves Learning from Noisily Labeled Long-Tailed Data [119.0612617460727]
Long-tailed and noisily labeled data frequently appear in real-world applications and impose significant challenges for learning.
We introduce the Fairness Regularizer (FR), inspired by regularizing the performance gap between any two sub-populations.
We show that the introduced fairness regularizer improves the performances of sub-populations on the tail and the overall learning performance.
arXiv Detail & Related papers (2023-03-22T03:46:51Z) - CLAMP: Prompt-based Contrastive Learning for Connecting Language and
Animal Pose [70.59906971581192]
We introduce a novel prompt-based Contrastive learning scheme for connecting Language and AniMal Pose effectively.
The CLAMP attempts to bridge the gap by adapting the text prompts to the animal keypoints during network training.
Experimental results show that our method achieves state-of-the-art performance under the supervised, few-shot, and zero-shot settings.
arXiv Detail & Related papers (2022-06-23T14:51:42Z) - Segmentation Enhanced Lameness Detection in Dairy Cows from RGB and
Depth Video [8.906235809404189]
Early lameness detection helps farmers address illnesses early and avoid negative effects caused by the degeneration of cows' condition.
We collected a dataset of short clips of cows exiting a milking station and annotated the degree of lameness of the cows.
We proposed a lameness detection method that leverages pre-trained neural networks to extract discriminative features from videos and assign a binary score to each cow indicating its condition: "healthy" or "lame"
arXiv Detail & Related papers (2022-06-09T12:16:31Z) - Self-supervised models of audio effectively explain human cortical
responses to speech [71.57870452667369]
We capitalize on the progress of self-supervised speech representation learning to create new state-of-the-art models of the human auditory system.
We show that these results show that self-supervised models effectively capture the hierarchy of information relevant to different stages of speech processing in human cortex.
arXiv Detail & Related papers (2022-05-27T22:04:02Z) - T-LEAP: occlusion-robust pose estimation of walking cows using temporal
information [0.0]
Lameness, a prevalent health disorder in dairy cows, is commonly detected by analyzing the gait of cows.
A cow's gait can be tracked in videos using pose estimation models because models learn to automatically localize anatomical landmarks in images and videos.
Most animal pose estimation models are static, that is, videos are processed frame by frame and do not use any temporal information.
arXiv Detail & Related papers (2021-04-16T10:50:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.