Seabed-Net: A multi-task network for joint bathymetry estimation and seabed classification from remote sensing imagery in shallow waters
- URL: http://arxiv.org/abs/2510.19329v1
- Date: Wed, 22 Oct 2025 07:43:03 GMT
- Title: Seabed-Net: A multi-task network for joint bathymetry estimation and seabed classification from remote sensing imagery in shallow waters
- Authors: Panagiotis Agrafiotis, Begüm Demir,
- Abstract summary: Seabed-Net is a unified multi-task framework that simultaneously predicts bathymetry and seabed classification.<n>It consistently outperforms traditional empirical models and machine learning regression methods.<n>It also reduces bathymetric RMSE by 10-30% compared to state-of-the-art single-task and multi-task baselines.
- Score: 6.101839518775969
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Accurate, detailed, and regularly updated bathymetry, coupled with complex semantic content, is essential for under-mapped shallow-water environments facing increasing climatological and anthropogenic pressures. However, existing approaches that derive either depth or seabed classes from remote sensing imagery treat these tasks in isolation, forfeiting the mutual benefits of their interaction and hindering the broader adoption of deep learning methods. To address these limitations, we introduce Seabed-Net, a unified multi-task framework that simultaneously predicts bathymetry and pixel-based seabed classification from remote sensing imagery of various resolutions. Seabed-Net employs dual-branch encoders for bathymetry estimation and pixel-based seabed classification, integrates cross-task features via an Attention Feature Fusion module and a windowed Swin-Transformer fusion block, and balances objectives through dynamic task uncertainty weighting. In extensive evaluations at two heterogeneous coastal sites, it consistently outperforms traditional empirical models and traditional machine learning regression methods, achieving up to 75\% lower RMSE. It also reduces bathymetric RMSE by 10-30\% compared to state-of-the-art single-task and multi-task baselines and improves seabed classification accuracy up to 8\%. Qualitative analyses further demonstrate enhanced spatial consistency, sharper habitat boundaries, and corrected depth biases in low-contrast regions. These results confirm that jointly modeling depth with both substrate and seabed habitats yields synergistic gains, offering a robust, open solution for integrated shallow-water mapping. Code and pretrained weights are available at https://github.com/pagraf/Seabed-Net.
Related papers
- High-Resolution Underwater Camouflaged Object Detection: GBU-UCOD Dataset and Topology-Aware and Frequency-Decoupled Networks [32.76569239634241]
We propose a novel framework that integrates topology-aware modeling with frequency-decoupled perception.<n>DeepTopo-Net achieves state-of-the-art performance, particularly in preserving morphological integrity of complex underwater patterns.
arXiv Detail & Related papers (2026-02-03T14:41:27Z) - UDPNet: Unleashing Depth-based Priors for Robust Image Dehazing [77.10640210751981]
UDPNet is a general framework that leverages depth-based priors from a large-scale pretrained depth estimation model DepthAnything V2.<n>Our proposed solution establishes a new benchmark for depth-aware dehazing across various scenarios.
arXiv Detail & Related papers (2026-01-11T13:29:02Z) - Scale-invariant and View-relational Representation Learning for Full Surround Monocular Depth [21.216686704928474]
We propose a knowledge distillation strategy that transfers robust depth knowledge from a foundation model to a lightweight FSMDE network.<n>Our method achieves a favorable trade-off between performance and efficiency, meeting real-time requirements.
arXiv Detail & Related papers (2025-12-09T15:17:20Z) - APGNet: Adaptive Prior-Guided for Underwater Camouflaged Object Detection [22.097955383220143]
We propose an Adaptive Prior-Guided Network (APGNet) to detect camouflaged objects in underwater environments.<n>APGNet integrates a Siamese architecture with a novel prior-guided mechanism to enhance robustness and detection accuracy.<n>Our proposed method APGNet outperforms 15 state-of-art methods under widely used evaluation metrics.
arXiv Detail & Related papers (2025-10-14T01:51:44Z) - Neptune-X: Active X-to-Maritime Generation for Universal Maritime Object Detection [54.1960918379255]
Neptune-X is a data-centric generative-selection framework for maritime object detection.<n>X-to-Maritime is a multi-modality-conditioned generative model that synthesizes diverse and realistic maritime scenes.<n>Our approach sets a new benchmark in maritime scene synthesis, significantly improving detection accuracy.
arXiv Detail & Related papers (2025-09-25T04:59:02Z) - Tree-Mamba: A Tree-Aware Mamba for Underwater Monocular Depth Estimation [85.17735565146106]
Underwater Monocular Depth Estimation (UMDE) is a critical task that aims to estimate high-precision depth maps from underwater degraded images.<n>We develop a novel tree-aware Mamba method, dubbed Tree-Mamba, for estimating accurate monocular depth maps from underwater degraded images.<n>We construct an underwater depth estimation benchmark (called BlueDepth), which consists of 38,162 underwater image pairs with reliable depth labels.
arXiv Detail & Related papers (2025-07-10T12:10:51Z) - YH-MINER: Multimodal Intelligent System for Natural Ecological Reef Metric Extraction [23.4289262373633]
Coral reefs, crucial for sustaining marine biodiversity and ecological processes, face escalating threats.<n>This study develops the YH-MINER system, establishing an intelligent framework for "object detection-semantic segmentation-prior input"<n>The system achieves genus-level classification accuracy of 88% and simultaneously extracting core ecological metrics.
arXiv Detail & Related papers (2025-05-28T11:36:18Z) - Occlusion Boundary and Depth: Mutual Enhancement via Multi-Task Learning [3.4174356345935393]
We propose MoDOT, a novel method that jointly estimates depth and OBs from a single image.<n>MoDOT incorporates a new module, CASM, which combines cross-attention and multi-scale strip convolutions to leverage mid-level OB features.<n>Experiments demonstrate the mutual benefits of jointly estimating depth and OBs, and validate the effectiveness of MoDOT's design.
arXiv Detail & Related papers (2025-05-27T14:15:19Z) - Relative Pose Estimation through Affine Corrections of Monocular Depth Priors [69.59216331861437]
We develop three solvers for relative pose estimation that explicitly account for independent affine (scale and shift) ambiguities.<n>We propose a hybrid estimation pipeline that combines our proposed solvers with classic point-based solvers and epipolar constraints.
arXiv Detail & Related papers (2025-01-09T18:58:30Z) - DepthSplat: Connecting Gaussian Splatting and Depth [90.06180236292866]
We present DepthSplat to connect Gaussian splatting and depth estimation.<n>We show that Gaussian splatting can serve as an unsupervised pre-training objective for learning powerful depth models.<n>Our DepthSplat achieves state-of-the-art performance on ScanNet, RealEstate10K and DL3DV datasets.
arXiv Detail & Related papers (2024-10-17T17:59:58Z) - GAM-Depth: Self-Supervised Indoor Depth Estimation Leveraging a
Gradient-Aware Mask and Semantic Constraints [12.426365333096264]
We propose GAM-Depth, developed upon two novel components: gradient-aware mask and semantic constraints.
The gradient-aware mask enables adaptive and robust supervision for both key areas and textureless regions.
The incorporation of semantic constraints for indoor self-supervised depth estimation improves depth discrepancies at object boundaries.
arXiv Detail & Related papers (2024-02-22T07:53:34Z) - PUGAN: Physical Model-Guided Underwater Image Enhancement Using GAN with
Dual-Discriminators [120.06891448820447]
How to obtain clear and visually pleasant images has become a common concern of people.
The task of underwater image enhancement (UIE) has also emerged as the times require.
In this paper, we propose a physical model-guided GAN model for UIE, referred to as PUGAN.
Our PUGAN outperforms state-of-the-art methods in both qualitative and quantitative metrics.
arXiv Detail & Related papers (2023-06-15T07:41:12Z) - DeepWSD: Projecting Degradations in Perceptual Space to Wasserstein
Distance in Deep Feature Space [67.07476542850566]
We propose to model the quality degradation in perceptual space from a statistical distribution perspective.
The quality is measured based upon the Wasserstein distance in the deep feature domain.
The deep Wasserstein distance (DeepWSD) performed on features from neural networks enjoys better interpretability of the quality contamination.
arXiv Detail & Related papers (2022-08-05T02:46:12Z) - SVAM: Saliency-guided Visual Attention Modeling by Autonomous Underwater
Robots [16.242924916178282]
This paper presents a holistic approach to saliency-guided visual attention modeling (SVAM) for use by autonomous underwater robots.
Our proposed model, named SVAM-Net, integrates deep visual features at various scales and semantics for effective salient object detection (SOD) in natural underwater images.
arXiv Detail & Related papers (2020-11-12T08:17:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.