Pig aggression classification using CNN, Transformers and Recurrent
Networks
- URL: http://arxiv.org/abs/2403.08528v1
- Date: Wed, 13 Mar 2024 13:38:58 GMT
- Title: Pig aggression classification using CNN, Transformers and Recurrent
Networks
- Authors: Junior Silva Souza, Eduardo Bedin, Gabriel Toshio Hirokawa Higa,
Newton Loebens, Hemerson Pistori
- Abstract summary: Aggressiveness in pigs is an example of behavior that is studied to reduce its impact through animal classification and identification.
The main techniques utilized in this study are variants of transformers: STAM, TimeSformer, and ViViT, as well as techniques using convolutions, such as ResNet3D2, Resnet(2+1)D, and CnnLstm.
- Score: 0.3792473194193801
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The development of techniques that can be used to analyze and detect animal
behavior is a crucial activity for the livestock sector, as it is possible to
monitor the stress and animal welfare and contributes to decision making in the
farm. Thus, the development of applications can assist breeders in making
decisions to improve production performance and reduce costs, once the animal
behavior is analyzed by humans and this can lead to susceptible errors and time
consumption. Aggressiveness in pigs is an example of behavior that is studied
to reduce its impact through animal classification and identification. However,
this process is laborious and susceptible to errors, which can be reduced
through automation by visually classifying videos captured in controlled
environment. The captured videos can be used for training and, as a result, for
classification through computer vision and artificial intelligence, employing
neural network techniques. The main techniques utilized in this study are
variants of transformers: STAM, TimeSformer, and ViViT, as well as techniques
using convolutions, such as ResNet3D2, Resnet(2+1)D, and CnnLstm. These
techniques were employed for pig video classification with the objective of
identifying aggressive and non-aggressive behaviors. In this work, various
techniques were compared to analyze the contribution of using transformers, in
addition to the effectiveness of the convolution technique in video
classification. The performance was evaluated using accuracy, precision, and
recall. The TimerSformer technique showed the best results in video
classification, with median accuracy of 0.729.
Related papers
- What Makes Pre-Trained Visual Representations Successful for Robust
Manipulation? [57.92924256181857]
We find that visual representations designed for manipulation and control tasks do not necessarily generalize under subtle changes in lighting and scene texture.
We find that emergent segmentation ability is a strong predictor of out-of-distribution generalization among ViT models.
arXiv Detail & Related papers (2023-11-03T18:09:08Z) - Defect Classification in Additive Manufacturing Using CNN-Based Vision
Processing [76.72662577101988]
This paper examines two scenarios: first, using convolutional neural networks (CNNs) to accurately classify defects in an image dataset from AM and second, applying active learning techniques to the developed classification model.
This allows the construction of a human-in-the-loop mechanism to reduce the size of the data required to train and generate training data.
arXiv Detail & Related papers (2023-07-14T14:36:58Z) - CVB: A Video Dataset of Cattle Visual Behaviors [13.233877352490923]
Existing datasets for cattle behavior recognition are mostly small, lack well-defined labels, or are collected in unrealistic controlled environments.
We introduce a new dataset, called Cattle Visual Behaviors (CVB), that consists of 502 video clips, each fifteen seconds long, captured in natural lighting conditions, and annotated with eleven visually perceptible behaviors of grazing cattle.
arXiv Detail & Related papers (2023-05-26T00:44:11Z) - CNN-Based Action Recognition and Pose Estimation for Classifying Animal
Behavior from Videos: A Survey [0.0]
Action recognition, classifying activities performed by one or more subjects in a trimmed video, forms the basis of many techniques.
Deep learning models for human action recognition have progressed over the last decade.
Recent interest in research that incorporates deep learning-based action recognition for classification has increased.
arXiv Detail & Related papers (2023-01-15T20:54:44Z) - TempNet: Temporal Attention Towards the Detection of Animal Behaviour in
Videos [63.85815474157357]
We propose an efficient computer vision- and deep learning-based method for the detection of biological behaviours in videos.
TempNet uses an encoder bridge and residual blocks to maintain model performance with a two-staged, spatial, then temporal, encoder.
We demonstrate its application to the detection of sablefish (Anoplopoma fimbria) startle events.
arXiv Detail & Related papers (2022-11-17T23:55:12Z) - Wild Animal Classifier Using CNN [0.0]
Convolution neural networks (CNNs) have multiple layers which have different weights for the purpose of prediction of a particular input.
Image segmentation is one such widely used image processing method which provides a clear demarcation of the areas of interest in the image.
arXiv Detail & Related papers (2022-10-03T13:14:08Z) - Revisiting Classifier: Transferring Vision-Language Models for Video
Recognition [102.93524173258487]
Transferring knowledge from task-agnostic pre-trained deep models for downstream tasks is an important topic in computer vision research.
In this study, we focus on transferring knowledge for video classification tasks.
We utilize the well-pretrained language model to generate good semantic target for efficient transferring learning.
arXiv Detail & Related papers (2022-07-04T10:00:47Z) - Ultrafast Image Categorization in Biology and Neural Models [0.0]
We re-trained the standard VGG 16 CNN on two independent tasks that are ecologically relevant to humans.
We show that re-training the network achieves a human-like level of performance, comparable to that reported in psychophysical tasks.
arXiv Detail & Related papers (2022-05-07T11:19:40Z) - Self-supervised Transformer for Deepfake Detection [112.81127845409002]
Deepfake techniques in real-world scenarios require stronger generalization abilities of face forgery detectors.
Inspired by transfer learning, neural networks pre-trained on other large-scale face-related tasks may provide useful features for deepfake detection.
In this paper, we propose a self-supervised transformer based audio-visual contrastive learning method.
arXiv Detail & Related papers (2022-03-02T17:44:40Z) - Overcoming the Domain Gap in Neural Action Representations [60.47807856873544]
3D pose data can now be reliably extracted from multi-view video sequences without manual intervention.
We propose to use it to guide the encoding of neural action representations together with a set of neural and behavioral augmentations.
To reduce the domain gap, during training, we swap neural and behavioral data across animals that seem to be performing similar actions.
arXiv Detail & Related papers (2021-12-02T12:45:46Z) - Dairy Cow rumination detection: A deep learning approach [0.8312466807725921]
Rumination behavior is a significant variable for tracking the development and yield of animal husbandry.
Modern attached devices are invasive, stressful and uncomfortable for the cattle.
In this study, we introduce an innovative monitoring method using Convolution Neural Network (CNN)-based deep learning models.
arXiv Detail & Related papers (2021-01-07T07:33:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.