Image Labels Are All You Need for Coarse Seagrass Segmentation
- URL: http://arxiv.org/abs/2303.00973v2
- Date: Wed, 6 Sep 2023 01:48:56 GMT
- Title: Image Labels Are All You Need for Coarse Seagrass Segmentation
- Authors: Scarlett Raine, Ross Marchant, Brano Kusy, Frederic Maire and Tobias
Fischer
- Abstract summary: Seagrass meadows serve as critical carbon sinks, but estimating the amount of carbon they store requires knowledge of the seagrass species present.
Previous approaches for seagrass detection and classification have required supervision from patch-level labels.
We introduce SeaFeats, an architecture that uses unsupervised contrastive pre-training and feature similarity, and SeaCLIP, a model that showcases the effectiveness of large language models as a supervisory signal.
- Score: 3.253176232272777
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Seagrass meadows serve as critical carbon sinks, but estimating the amount of
carbon they store requires knowledge of the seagrass species present.
Underwater and surface vehicles equipped with machine learning algorithms can
help to accurately estimate the composition and extent of seagrass meadows at
scale. However, previous approaches for seagrass detection and classification
have required supervision from patch-level labels. In this paper, we reframe
seagrass classification as a weakly supervised coarse segmentation problem
where image-level labels are used during training (25 times fewer labels
compared to patch-level labeling) and patch-level outputs are obtained at
inference time. To this end, we introduce SeaFeats, an architecture that uses
unsupervised contrastive pre-training and feature similarity, and SeaCLIP, a
model that showcases the effectiveness of large language models as a
supervisory signal in domain-specific applications. We demonstrate that an
ensemble of SeaFeats and SeaCLIP leads to highly robust performance. Our method
outperforms previous approaches that require patch-level labels on the
multi-species 'DeepSeagrass' dataset by 6.8% (absolute) for the class-weighted
F1 score, and by 12.1% (absolute) for the seagrass presence/absence F1 score on
the 'Global Wetlands' dataset. We also present two case studies for real-world
deployment: outlier detection on the Global Wetlands dataset, and application
of our method on imagery collected by the FloatyBoat autonomous surface
vehicle.
Related papers
- Co-training for Low Resource Scientific Natural Language Inference [65.37685198688538]
We propose a novel co-training method that assigns weights based on the training dynamics of the classifiers to the distantly supervised labels.
By assigning importance weights instead of filtering out examples based on an arbitrary threshold on the predicted confidence, we maximize the usage of automatically labeled data.
The proposed method obtains an improvement of 1.5% in Macro F1 over the distant supervision baseline, and substantial improvements over several other strong SSL baselines.
arXiv Detail & Related papers (2024-06-20T18:35:47Z) - LayerMatch: Do Pseudo-labels Benefit All Layers? [77.59625180366115]
Semi-supervised learning offers a promising solution to mitigate the dependency of labeled data.
We develop two layer-specific pseudo-label strategies, termed Grad-ReLU and Avg-Clustering.
Our approach consistently demonstrates exceptional performance on standard semi-supervised learning benchmarks.
arXiv Detail & Related papers (2024-06-20T11:25:50Z) - Deep learning for multi-label classification of coral conditions in the
Indo-Pacific via underwater photogrammetry [24.00646413446011]
This study created a dataset representing common coral conditions and associated stressors in the Indo-Pacific.
It assessed existing classification algorithms and proposed a new multi-label method for automatically detecting coral conditions and extracting ecological information.
The proposed method accurately classified coral conditions as healthy, compromised, dead, and rubble.
arXiv Detail & Related papers (2024-03-09T14:42:16Z) - WhaleNet: a Novel Deep Learning Architecture for Marine Mammals Vocalizations on Watkins Marine Mammal Sound Database [49.1574468325115]
We introduce textbfWhaleNet (Wavelet Highly Adaptive Learning Ensemble Network), a sophisticated deep ensemble architecture for the classification of marine mammal vocalizations.
We achieve an improvement in classification accuracy by $8-10%$ over existing architectures, corresponding to a classification accuracy of $97.61%$.
arXiv Detail & Related papers (2024-02-20T11:36:23Z) - Pseudo-label Alignment for Semi-supervised Instance Segmentation [67.9616087910363]
Pseudo-labeling is significant for semi-supervised instance segmentation.
In existing pipelines, pseudo-labels that contain valuable information may be filtered out due to mismatches in class and mask quality.
We propose a novel framework, called pseudo-label aligning instance segmentation (PAIS), in this paper.
arXiv Detail & Related papers (2023-08-10T05:56:53Z) - Enhancing Self-Supervised Learning for Remote Sensing with Elevation
Data: A Case Study with Scarce And High Level Semantic Labels [1.534667887016089]
This work proposes a hybrid unsupervised and supervised learning method to pre-train models applied in Earth observation downstream tasks.
We combine a contrastive approach to pre-train models with a pixel-wise regression pre-text task to predict coarse elevation maps.
arXiv Detail & Related papers (2023-04-13T23:01:11Z) - Habitat classification from satellite observations with sparse
annotations [4.164845768197488]
We propose a method for habitat classification and mapping using remote sensing data.
The method is characterized by using finely-grained, sparse, single-pixel annotations collected from the field.
We show that cropping augmentations, test-time augmentation and semi-supervised learning can help classification even further.
arXiv Detail & Related papers (2022-09-26T20:14:59Z) - Rethinking Pseudo Labels for Semi-Supervised Object Detection [84.697097472401]
We introduce certainty-aware pseudo labels tailored for object detection.
We dynamically adjust the thresholds used to generate pseudo labels and reweight loss functions for each category to alleviate the class imbalance problem.
Our approach improves supervised baselines by up to 10% AP using only 1-10% labeled data from COCO.
arXiv Detail & Related papers (2021-06-01T01:32:03Z) - Highly Efficient Representation and Active Learning Framework for
Imbalanced Data and its Application to COVID-19 X-Ray Classification [0.7829352305480284]
We propose a highly data-efficient classification and active learning framework for classifying chest X-rays.
It is based on (1) unsupervised representation learning of a Convolutional Neural Network and (2) the Gaussian Process method.
We demonstrate that only $sim 10%$ of the labeled data is needed to reach the accuracy from training all available labels.
arXiv Detail & Related papers (2021-02-25T02:48:59Z) - Joint Visual and Temporal Consistency for Unsupervised Domain Adaptive
Person Re-Identification [64.37745443119942]
This paper jointly enforces visual and temporal consistency in the combination of a local one-hot classification and a global multi-class classification.
Experimental results on three large-scale ReID datasets demonstrate the superiority of proposed method in both unsupervised and unsupervised domain adaptive ReID tasks.
arXiv Detail & Related papers (2020-07-21T14:31:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.