CoralVOS: Dataset and Benchmark for Coral Video Segmentation
- URL: http://arxiv.org/abs/2310.01946v1
- Date: Tue, 3 Oct 2023 10:45:37 GMT
- Title: CoralVOS: Dataset and Benchmark for Coral Video Segmentation
- Authors: Zheng Ziqiang, Xie Yaofeng, Liang Haixin, Yu Zhibin, Sai-Kit Yeung
- Abstract summary: We propose a large-scale coral video segmentation dataset: textbfCoralVOS as demonstrated in Fig. 1.
We perform experiments on our CoralVOS dataset, including 6 recent state-of-the-art video object segmentation (VOS) algorithms.
The results show that there is still great potential for further promoting the segmentation accuracy.
- Score: 12.434773034255455
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Coral reefs formulate the most valuable and productive marine ecosystems,
providing habitat for many marine species. Coral reef surveying and analysis
are currently confined to coral experts who invest substantial effort in
generating comprehensive and dependable reports (\emph{e.g.}, coral coverage,
population, spatial distribution, \textit{etc}), from the collected survey
data. However, performing dense coral analysis based on manual efforts is
significantly time-consuming, the existing coral analysis algorithms compromise
and opt for performing down-sampling and only conducting sparse point-based
coral analysis within selected frames. However, such down-sampling will
\textbf{inevitable} introduce the estimation bias or even lead to wrong
results. To address this issue, we propose to perform \textbf{dense coral video
segmentation}, with no down-sampling involved. Through video object
segmentation, we could generate more \textit{reliable} and \textit{in-depth}
coral analysis than the existing coral reef analysis algorithms. To boost such
dense coral analysis, we propose a large-scale coral video segmentation
dataset: \textbf{CoralVOS} as demonstrated in Fig. 1. To the best of our
knowledge, our CoralVOS is the first dataset and benchmark supporting dense
coral video segmentation. We perform experiments on our CoralVOS dataset,
including 6 recent state-of-the-art video object segmentation (VOS) algorithms.
We fine-tuned these VOS algorithms on our CoralVOS dataset and achieved
observable performance improvement. The results show that there is still great
potential for further promoting the segmentation accuracy. The dataset and
trained models will be released with the acceptance of this work to foster the
coral reef research community.
Related papers
- Automatic Coral Detection with YOLO: A Deep Learning Approach for Efficient and Accurate Coral Reef Monitoring [0.0]
Coral reefs are vital ecosystems that are under increasing threat due to local human impacts and climate change.
In this paper, we present an automatic coral detection system utilizing the You Only Look Once deep learning model.
arXiv Detail & Related papers (2024-04-03T08:00:46Z) - Deep learning for multi-label classification of coral conditions in the
Indo-Pacific via underwater photogrammetry [24.00646413446011]
This study created a dataset representing common coral conditions and associated stressors in the Indo-Pacific.
It assessed existing classification algorithms and proposed a new multi-label method for automatically detecting coral conditions and extracting ecological information.
The proposed method accurately classified coral conditions as healthy, compromised, dead, and rubble.
arXiv Detail & Related papers (2024-03-09T14:42:16Z) - Scalable Semantic 3D Mapping of Coral Reefs with Deep Learning [4.8902950939676675]
This paper presents a new paradigm for mapping underwater environments from ego-motion video.
We show high-precision 3D semantic mapping at unprecedented scale with significantly reduced required labor costs.
Our approach significantly scales up coral reef monitoring by taking a leap towards fully automatic analysis of video transects.
arXiv Detail & Related papers (2023-09-22T11:35:10Z) - Pengembangan Model untuk Mendeteksi Kerusakan pada Terumbu Karang dengan
Klasifikasi Citra [3.254879465902239]
This study utilizes a specialized dataset consisting of 923 images collected from Flickr using the Flickr API.
The method employed in this research involves the use of machine learning models, particularly convolutional neural networks (CNN)
It was found that a from-scratch ResNet model can outperform pretrained models in terms of precision and accuracy.
arXiv Detail & Related papers (2023-08-08T15:30:08Z) - Reef-insight: A framework for reef habitat mapping with clustering
methods via remote sensing [0.3670422696827526]
We present Reef-Insight, an unsupervised machine learning framework that features advanced clustering methods and remote sensing for reef habitat mapping.
Our framework compares different clustering methods for reef habitat mapping using remote sensing data.
Our results indicate that Reef-Insight can generate detailed reef habitat maps outlining distinct reef habitats.
arXiv Detail & Related papers (2023-01-26T00:03:09Z) - Efficient Per-Shot Convex Hull Prediction By Recurrent Learning [50.94452824380868]
We propose a deep learning based method of content aware convex hull prediction.
We employ a recurrent convolutional network (RCN) to implicitly analyze the complexity of video shots in order to predict their convex hulls.
Our experimental results reveal that our proposed model better approximations of the optimal convex hulls, and offers competitive time savings as compared to existing approaches.
arXiv Detail & Related papers (2022-06-10T05:11:02Z) - Video-based Person Re-identification without Bells and Whistles [49.51670583977911]
Video-based person re-identification (Re-ID) aims at matching the video tracklets with cropped video frames for identifying the pedestrians under different cameras.
There exists severe spatial and temporal misalignment for those cropped tracklets due to the imperfect detection and tracking results generated with obsolete methods.
We present a simple re-Detect and Link (DL) module which can effectively reduce those unexpected noise through applying the deep learning-based detection and tracking on the cropped tracklets.
arXiv Detail & Related papers (2021-05-22T10:17:38Z) - Deep ensembles based on Stochastic Activation Selection for Polyp
Segmentation [82.61182037130406]
This work deals with medical image segmentation and in particular with accurate polyp detection and segmentation during colonoscopy examinations.
Basic architecture in image segmentation consists of an encoder and a decoder.
We compare some variant of the DeepLab architecture obtained by varying the decoder backbone.
arXiv Detail & Related papers (2021-04-02T02:07:37Z) - Channelized Axial Attention for Semantic Segmentation [70.14921019774793]
We propose the Channelized Axial Attention (CAA) to seamlessly integratechannel attention and axial attention with reduced computationalcomplexity.
Our CAA not onlyrequires much less computation resources compared with otherdual attention models such as DANet, but also outperforms the state-of-the-art ResNet-101-based segmentation models on alltested datasets.
arXiv Detail & Related papers (2021-01-19T03:08:03Z) - The Devil is in Classification: A Simple Framework for Long-tail Object
Detection and Instance Segmentation [93.17367076148348]
We investigate performance drop of the state-of-the-art two-stage instance segmentation model Mask R-CNN on the recent long-tail LVIS dataset.
We unveil that a major cause is the inaccurate classification of object proposals.
We propose a simple calibration framework to more effectively alleviate classification head bias with a bi-level class balanced sampling approach.
arXiv Detail & Related papers (2020-07-23T12:49:07Z) - DenoiSeg: Joint Denoising and Segmentation [75.91760529986958]
We propose DenoiSeg, a new method that can be trained end-to-end on only a few annotated ground truth segmentations.
We achieve this by extending Noise2Void, a self-supervised denoising scheme that can be trained on noisy images alone, to also predict dense 3-class segmentations.
arXiv Detail & Related papers (2020-05-06T17:42:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.