Towards Self-Supervision for Video Identification of Individual
Holstein-Friesian Cattle: The Cows2021 Dataset
- URL: http://arxiv.org/abs/2105.01938v1
- Date: Wed, 5 May 2021 09:08:19 GMT
- Title: Towards Self-Supervision for Video Identification of Individual
Holstein-Friesian Cattle: The Cows2021 Dataset
- Authors: Jing Gao, Tilo Burghardt, William Andrew, Andrew W. Dowsey, Neill W.
Campbell
- Abstract summary: We publish the largest identity-annotated Holstein-Friesian cattle dataset Cows2021.
We propose exploiting the temporal coat pattern appearance across videos as a self-supervision signal for animal identity learning.
Results show an accuracy of Top-1 57.0% and Top-4: 76.9% and an Adjusted Rand Index: 0.53 compared to the ground truth.
- Score: 10.698921107213554
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In this paper we publish the largest identity-annotated Holstein-Friesian
cattle dataset Cows2021 and a first self-supervision framework for video
identification of individual animals. The dataset contains 10,402 RGB images
with labels for localisation and identity as well as 301 videos from the same
herd. The data shows top-down in-barn imagery, which captures the breed's
individually distinctive black and white coat pattern. Motivated by the
labelling burden involved in constructing visual cattle identification systems,
we propose exploiting the temporal coat pattern appearance across videos as a
self-supervision signal for animal identity learning. Using an
individual-agnostic cattle detector that yields oriented bounding-boxes,
rotation-normalised tracklets of individuals are formed via
tracking-by-detection and enriched via augmentations. This produces a
`positive' sample set per tracklet, which is paired against a `negative' set
sampled from random cattle of other videos. Frame-triplet contrastive learning
is then employed to construct a metric latent space. The fitting of a Gaussian
Mixture Model to this space yields a cattle identity classifier. Results show
an accuracy of Top-1 57.0% and Top-4: 76.9% and an Adjusted Rand Index: 0.53
compared to the ground truth. Whilst supervised training surpasses this
benchmark by a large margin, we conclude that self-supervision can nevertheless
play a highly effective role in speeding up labelling efforts when initially
constructing supervision information. We provide all data and full source code
alongside an analysis and evaluation of the system.
Related papers
- MultiCamCows2024 -- A Multi-view Image Dataset for AI-driven Holstein-Friesian Cattle Re-Identification on a Working Farm [2.9391768712283772]
We present MultiCamCows2024, a farm-scale image dataset filmed across multiple cameras for the biometric identification of individual Holstein-Friesian cattle.
The dataset comprises 101, 329 images of 90 cows, plus the underlying original CCTV footage.
We show that our framework enables fully automatic cattle identification, barring only the simple human verification of tracklet integrity.
arXiv Detail & Related papers (2024-10-16T15:58:47Z) - CNN Based Flank Predictor for Quadruped Animal Species [1.502956022927019]
We train a flank predictor that predicts the visible flank of quadruped mammalian species in images.
The developed models were evaluated in different scenarios of different unknown quadruped species in known and unknown environments.
The best model, trained on an EfficientNetV2 backbone, achieved an accuracy of 88.70 % for the unknown species lynx in a complex habitat.
arXiv Detail & Related papers (2024-06-19T14:24:26Z) - Universal Bovine Identification via Depth Data and Deep Metric Learning [1.6605913858547239]
This paper proposes and evaluates, for the first time, a depth-only deep learning system for accurately identifying individual cattle.
An increase in herd size skews the cow-to-human ratio at the farm and makes the manual monitoring of individuals more challenging.
Underpinned by our previous work, this paper introduces a deep-metric learning method for cattle identification using depth data from an off-the-shelf 3D camera.
arXiv Detail & Related papers (2024-03-29T22:03:53Z) - APT-36K: A Large-scale Benchmark for Animal Pose Estimation and Tracking [77.87449881852062]
APT-36K is the first large-scale benchmark for animal pose estimation and tracking.
It consists of 2,400 video clips collected and filtered from 30 animal species with 15 frames for each video, resulting in 36,000 frames in total.
We benchmark several representative models on the following three tracks: (1) supervised animal pose estimation on a single frame under intra- and inter-domain transfer learning settings, (2) inter-species domain generalization test for unseen animals, and (3) animal pose estimation with animal tracking.
arXiv Detail & Related papers (2022-06-12T07:18:36Z) - Label a Herd in Minutes: Individual Holstein-Friesian Cattle
Identification [12.493458478953515]
We describe a practically evaluated approach for training visual cattle ID systems for a whole farm requiring only ten minutes of labelling effort.
For the task of automatic identification of individual Holstein-Friesians in real-world farm CCTV, we show that self-supervision, metric learning, cluster analysis, and active learning can complement each other.
arXiv Detail & Related papers (2022-04-22T19:41:47Z) - Persistent Animal Identification Leveraging Non-Visual Markers [71.14999745312626]
We aim to locate and provide a unique identifier for each mouse in a cluttered home-cage environment through time.
This is a very challenging problem due to (i) the lack of distinguishing visual features for each mouse, and (ii) the close confines of the scene with constant occlusion.
Our approach achieves 77% accuracy on this animal identification problem, and is able to reject spurious detections when the animals are hidden.
arXiv Detail & Related papers (2021-12-13T17:11:32Z) - ASCNet: Self-supervised Video Representation Learning with
Appearance-Speed Consistency [62.38914747727636]
We study self-supervised video representation learning, which is a challenging task due to 1) a lack of labels for explicit supervision and 2) unstructured and noisy visual information.
Existing methods mainly use contrastive loss with video clips as the instances and learn visual representation by discriminating instances from each other.
In this paper, we observe that the consistency between positive samples is the key to learn robust video representations.
arXiv Detail & Related papers (2021-06-04T08:44:50Z) - Pretrained equivariant features improve unsupervised landmark discovery [69.02115180674885]
We formulate a two-step unsupervised approach that overcomes this challenge by first learning powerful pixel-based features.
Our method produces state-of-the-art results in several challenging landmark detection datasets.
arXiv Detail & Related papers (2021-04-07T05:42:11Z) - Unsupervised Pretraining for Object Detection by Patch Reidentification [72.75287435882798]
Unsupervised representation learning achieves promising performances in pre-training representations for object detectors.
This work proposes a simple yet effective representation learning method for object detection, named patch re-identification (Re-ID)
Our method significantly outperforms its counterparts on COCO in all settings, such as different training iterations and data percentages.
arXiv Detail & Related papers (2021-03-08T15:13:59Z) - Labelling unlabelled videos from scratch with multi-modal
self-supervision [82.60652426371936]
unsupervised labelling of a video dataset does not come for free from strong feature encoders.
We propose a novel clustering method that allows pseudo-labelling of a video dataset without any human annotations.
An extensive analysis shows that the resulting clusters have high semantic overlap to ground truth human labels.
arXiv Detail & Related papers (2020-06-24T12:28:17Z) - Visual Identification of Individual Holstein-Friesian Cattle via Deep
Metric Learning [8.784100314325395]
Holstein-Friesian cattle exhibit individually-characteristic black and white coat patterns visually akin to those arising from Turing's reaction-diffusion systems.
This work takes advantage of these natural markings in order to automate visual detection and biometric identification of individual Holstein-Friesians via convolutional neural networks and deep metric learning techniques.
arXiv Detail & Related papers (2020-06-16T14:41:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.