CoralVOS: Dataset and Benchmark for Coral Video Segmentation
- URL: http://arxiv.org/abs/2310.01946v1
- Date: Tue, 3 Oct 2023 10:45:37 GMT
- Title: CoralVOS: Dataset and Benchmark for Coral Video Segmentation
- Authors: Zheng Ziqiang, Xie Yaofeng, Liang Haixin, Yu Zhibin, Sai-Kit Yeung
- Abstract summary: We propose a large-scale coral video segmentation dataset: textbfCoralVOS as demonstrated in Fig. 1.
We perform experiments on our CoralVOS dataset, including 6 recent state-of-the-art video object segmentation (VOS) algorithms.
The results show that there is still great potential for further promoting the segmentation accuracy.
- Score: 12.434773034255455
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Coral reefs formulate the most valuable and productive marine ecosystems,
providing habitat for many marine species. Coral reef surveying and analysis
are currently confined to coral experts who invest substantial effort in
generating comprehensive and dependable reports (\emph{e.g.}, coral coverage,
population, spatial distribution, \textit{etc}), from the collected survey
data. However, performing dense coral analysis based on manual efforts is
significantly time-consuming, the existing coral analysis algorithms compromise
and opt for performing down-sampling and only conducting sparse point-based
coral analysis within selected frames. However, such down-sampling will
\textbf{inevitable} introduce the estimation bias or even lead to wrong
results. To address this issue, we propose to perform \textbf{dense coral video
segmentation}, with no down-sampling involved. Through video object
segmentation, we could generate more \textit{reliable} and \textit{in-depth}
coral analysis than the existing coral reef analysis algorithms. To boost such
dense coral analysis, we propose a large-scale coral video segmentation
dataset: \textbf{CoralVOS} as demonstrated in Fig. 1. To the best of our
knowledge, our CoralVOS is the first dataset and benchmark supporting dense
coral video segmentation. We perform experiments on our CoralVOS dataset,
including 6 recent state-of-the-art video object segmentation (VOS) algorithms.
We fine-tuned these VOS algorithms on our CoralVOS dataset and achieved
observable performance improvement. The results show that there is still great
potential for further promoting the segmentation accuracy. The dataset and
trained models will be released with the acceptance of this work to foster the
coral reef research community.
Related papers
- CoralSCOP-LAT: Labeling and Analyzing Tool for Coral Reef Images with Dense Mask [14.092526875441221]
We propose CoralSCOP-LAT, an automatic and semi-automatic coral reef labeling and analysis tool.
The proposed CoralSCOP-LAT surpasses the existing tools by a large margin from analysis efficiency, accuracy, and flexibility.
Our CoralSCOP-LAT, as the first dense coral reef analysis tool in the market, facilitates repeated large-scale coral reef monitoring analysis.
arXiv Detail & Related papers (2024-10-27T13:26:44Z) - Automatic Coral Detection with YOLO: A Deep Learning Approach for Efficient and Accurate Coral Reef Monitoring [0.0]
Coral reefs are vital ecosystems that are under increasing threat due to local human impacts and climate change.
In this paper, we present an automatic coral detection system utilizing the You Only Look Once deep learning model.
arXiv Detail & Related papers (2024-04-03T08:00:46Z) - Deep learning for multi-label classification of coral conditions in the
Indo-Pacific via underwater photogrammetry [24.00646413446011]
This study created a dataset representing common coral conditions and associated stressors in the Indo-Pacific.
It assessed existing classification algorithms and proposed a new multi-label method for automatically detecting coral conditions and extracting ecological information.
The proposed method accurately classified coral conditions as healthy, compromised, dead, and rubble.
arXiv Detail & Related papers (2024-03-09T14:42:16Z) - Pengembangan Model untuk Mendeteksi Kerusakan pada Terumbu Karang dengan
Klasifikasi Citra [3.254879465902239]
This study utilizes a specialized dataset consisting of 923 images collected from Flickr using the Flickr API.
The method employed in this research involves the use of machine learning models, particularly convolutional neural networks (CNN)
It was found that a from-scratch ResNet model can outperform pretrained models in terms of precision and accuracy.
arXiv Detail & Related papers (2023-08-08T15:30:08Z) - Reef-insight: A framework for reef habitat mapping with clustering
methods via remote sensing [0.3670422696827526]
We present Reef-Insight, an unsupervised machine learning framework that features advanced clustering methods and remote sensing for reef habitat mapping.
Our framework compares different clustering methods for reef habitat mapping using remote sensing data.
Our results indicate that Reef-Insight can generate detailed reef habitat maps outlining distinct reef habitats.
arXiv Detail & Related papers (2023-01-26T00:03:09Z) - Video-based Person Re-identification without Bells and Whistles [49.51670583977911]
Video-based person re-identification (Re-ID) aims at matching the video tracklets with cropped video frames for identifying the pedestrians under different cameras.
There exists severe spatial and temporal misalignment for those cropped tracklets due to the imperfect detection and tracking results generated with obsolete methods.
We present a simple re-Detect and Link (DL) module which can effectively reduce those unexpected noise through applying the deep learning-based detection and tracking on the cropped tracklets.
arXiv Detail & Related papers (2021-05-22T10:17:38Z) - Deep ensembles based on Stochastic Activation Selection for Polyp
Segmentation [82.61182037130406]
This work deals with medical image segmentation and in particular with accurate polyp detection and segmentation during colonoscopy examinations.
Basic architecture in image segmentation consists of an encoder and a decoder.
We compare some variant of the DeepLab architecture obtained by varying the decoder backbone.
arXiv Detail & Related papers (2021-04-02T02:07:37Z) - Channelized Axial Attention for Semantic Segmentation [70.14921019774793]
We propose the Channelized Axial Attention (CAA) to seamlessly integratechannel attention and axial attention with reduced computationalcomplexity.
Our CAA not onlyrequires much less computation resources compared with otherdual attention models such as DANet, but also outperforms the state-of-the-art ResNet-101-based segmentation models on alltested datasets.
arXiv Detail & Related papers (2021-01-19T03:08:03Z) - Improving Generative Adversarial Networks with Local Coordinate Coding [150.24880482480455]
Generative adversarial networks (GANs) have shown remarkable success in generating realistic data from some predefined prior distribution.
In practice, semantic information might be represented by some latent distribution learned from data.
We propose an LCCGAN model with local coordinate coding (LCC) to improve the performance of generating data.
arXiv Detail & Related papers (2020-07-28T09:17:50Z) - The Devil is in Classification: A Simple Framework for Long-tail Object
Detection and Instance Segmentation [93.17367076148348]
We investigate performance drop of the state-of-the-art two-stage instance segmentation model Mask R-CNN on the recent long-tail LVIS dataset.
We unveil that a major cause is the inaccurate classification of object proposals.
We propose a simple calibration framework to more effectively alleviate classification head bias with a bi-level class balanced sampling approach.
arXiv Detail & Related papers (2020-07-23T12:49:07Z) - DenoiSeg: Joint Denoising and Segmentation [75.91760529986958]
We propose DenoiSeg, a new method that can be trained end-to-end on only a few annotated ground truth segmentations.
We achieve this by extending Noise2Void, a self-supervised denoising scheme that can be trained on noisy images alone, to also predict dense 3-class segmentations.
arXiv Detail & Related papers (2020-05-06T17:42:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.