Temperate Fish Detection and Classification: a Deep Learning based
Approach
- URL: http://arxiv.org/abs/2005.07518v1
- Date: Thu, 14 May 2020 12:40:57 GMT
- Title: Temperate Fish Detection and Classification: a Deep Learning based
Approach
- Authors: Kristian Muri Knausg{\aa}rd, Arne Wiklund, Tonje Knutsen S{\o}rdalen,
Kim Halvorsen, Alf Ring Kleiven, Lei Jiao, Morten Goodwin
- Abstract summary: We propose a two-step deep learning approach for the detection and classification of temperate fishes without pre-filtering.
The first step is to detect each single fish in an image, independent of species and sex.
In the second step, we adopt a Convolutional Neural Network (CNN) with the Squeeze-and-Excitation (SE) architecture for classifying each fish in the image without pre-filtering.
- Score: 6.282069822653608
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A wide range of applications in marine ecology extensively uses underwater
cameras. Still, to efficiently process the vast amount of data generated, we
need to develop tools that can automatically detect and recognize species
captured on film. Classifying fish species from videos and images in natural
environments can be challenging because of noise and variation in illumination
and the surrounding habitat. In this paper, we propose a two-step deep learning
approach for the detection and classification of temperate fishes without
pre-filtering. The first step is to detect each single fish in an image,
independent of species and sex. For this purpose, we employ the You Only Look
Once (YOLO) object detection technique. In the second step, we adopt a
Convolutional Neural Network (CNN) with the Squeeze-and-Excitation (SE)
architecture for classifying each fish in the image without pre-filtering. We
apply transfer learning to overcome the limited training samples of temperate
fishes and to improve the accuracy of the classification. This is done by
training the object detection model with ImageNet and the fish classifier via a
public dataset (Fish4Knowledge), whereupon both the object detection and
classifier are updated with temperate fishes of interest. The weights obtained
from pre-training are applied to post-training as a priori. Our solution
achieves the state-of-the-art accuracy of 99.27\% on the pre-training. The
percentage values for accuracy on the post-training are good; 83.68\% and
87.74\% with and without image augmentation, respectively, indicating that the
solution is viable with a more extensive dataset.
Related papers
- Improved detection of discarded fish species through BoxAL active learning [0.2544632696242629]
In this study, we present an active learning technique, named BoxAL, which includes estimation of epistemic certainty of the Faster R-CNN object-detection model.
The method allows selecting the most uncertain training images from an unlabeled pool, which are then used to train the object-detection model.
Our study additionally showed that the sampled new data is more valuable for training than the remaining unlabeled data.
arXiv Detail & Related papers (2024-10-07T10:01:30Z) - Automatic Coral Detection with YOLO: A Deep Learning Approach for Efficient and Accurate Coral Reef Monitoring [0.0]
Coral reefs are vital ecosystems that are under increasing threat due to local human impacts and climate change.
In this paper, we present an automatic coral detection system utilizing the You Only Look Once deep learning model.
arXiv Detail & Related papers (2024-04-03T08:00:46Z) - Rethinking Transformers Pre-training for Multi-Spectral Satellite
Imagery [78.43828998065071]
Recent advances in unsupervised learning have demonstrated the ability of large vision models to achieve promising results on downstream tasks.
Such pre-training techniques have also been explored recently in the remote sensing domain due to the availability of large amount of unlabelled data.
In this paper, we re-visit transformers pre-training and leverage multi-scale information that is effectively utilized with multiple modalities.
arXiv Detail & Related papers (2024-03-08T16:18:04Z) - Leveraging the Third Dimension in Contrastive Learning [88.17394309208925]
Self-Supervised Learning (SSL) methods operate on unlabeled data to learn robust representations useful for downstream tasks.
These augmentations ignore the fact that biological vision takes place in an immersive three-dimensional, temporally contiguous environment.
We explore two distinct approaches to incorporating depth signals into the SSL framework.
arXiv Detail & Related papers (2023-01-27T15:45:03Z) - TempNet: Temporal Attention Towards the Detection of Animal Behaviour in
Videos [63.85815474157357]
We propose an efficient computer vision- and deep learning-based method for the detection of biological behaviours in videos.
TempNet uses an encoder bridge and residual blocks to maintain model performance with a two-staged, spatial, then temporal, encoder.
We demonstrate its application to the detection of sablefish (Anoplopoma fimbria) startle events.
arXiv Detail & Related papers (2022-11-17T23:55:12Z) - Intelligent Masking: Deep Q-Learning for Context Encoding in Medical
Image Analysis [48.02011627390706]
We develop a novel self-supervised approach that occludes targeted regions to improve the pre-training procedure.
We show that training the agent against the prediction model can significantly improve the semantic features extracted for downstream classification tasks.
arXiv Detail & Related papers (2022-03-25T19:05:06Z) - A deep neural network for multi-species fish detection using multiple
acoustic cameras [0.0]
We present a novel approach that takes advantage of both CNN (Convolutional Neural Network) and classical CV (Computer Vision) techniques.
The pipeline pre-treats the acoustic images to extract 2 features, in order to localise the signals and improve the detection performances.
The YOLOv3-based model was trained with data of fish from multiple species recorded by the two common acoustic cameras.
arXiv Detail & Related papers (2021-09-22T11:47:24Z) - Rectifying the Shortcut Learning of Background: Shared Object
Concentration for Few-Shot Image Recognition [101.59989523028264]
Few-Shot image classification aims to utilize pretrained knowledge learned from a large-scale dataset to tackle a series of downstream classification tasks.
We propose COSOC, a novel Few-Shot Learning framework, to automatically figure out foreground objects at both pretraining and evaluation stage.
arXiv Detail & Related papers (2021-07-16T07:46:41Z) - Deep learning with self-supervision and uncertainty regularization to
count fish in underwater images [28.261323753321328]
Effective conservation actions require effective population monitoring.
Monitoring populations through image sampling has made data collection cheaper, wide-reaching and less intrusive.
Counting animals from such data is challenging, particularly when densely packed in noisy images.
Deep learning is the state-of-the-art method for many computer vision tasks, but it has yet to be properly explored to count animals.
arXiv Detail & Related papers (2021-04-30T13:02:19Z) - A Realistic Fish-Habitat Dataset to Evaluate Algorithms for Underwater
Visual Analysis [2.6476746128312194]
We present DeepFish as a benchmark suite with a large-scale dataset to train and test methods for several computer vision tasks.
The dataset consists of approximately 40 thousand images collected underwater from 20 greenhabitats in the marine-environments of tropical Australia.
Our experiments provide an in-depth analysis of the dataset characteristics, and the performance evaluation of several state-of-the-art approaches.
arXiv Detail & Related papers (2020-08-28T12:20:59Z) - Automatic image-based identification and biomass estimation of
invertebrates [70.08255822611812]
Time-consuming sorting and identification of taxa pose strong limitations on how many insect samples can be processed.
We propose to replace the standard manual approach of human expert-based sorting and identification with an automatic image-based technology.
We use state-of-the-art Resnet-50 and InceptionV3 CNNs for the classification task.
arXiv Detail & Related papers (2020-02-05T21:38:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.