Modality-Guided Subnetwork for Salient Object Detection
- URL: http://arxiv.org/abs/2110.04904v1
- Date: Sun, 10 Oct 2021 20:59:11 GMT
- Title: Modality-Guided Subnetwork for Salient Object Detection
- Authors: Zongwei Wu, Guillaume Allibert, Christophe Stolz, Chao Ma, C\'edric
Demonceaux
- Abstract summary: Most RGBD networks require multi-modalities from the input side and feed them separately through a two-stream design.
We present in this paper a novel fusion design named modality-guided subnetwork (MGSnet)
It has the following superior designs: 1) Our model works for both RGB and RGBD data, and dynamically estimating depth if not available.
- Score: 5.491692465987937
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent RGBD-based models for saliency detection have attracted research
attention. The depth clues such as boundary clues, surface normal, shape
attribute, etc., contribute to the identification of salient objects with
complicated scenarios. However, most RGBD networks require multi-modalities
from the input side and feed them separately through a two-stream design, which
inevitably results in extra costs on depth sensors and computation. To tackle
these inconveniences, we present in this paper a novel fusion design named
modality-guided subnetwork (MGSnet). It has the following superior designs: 1)
Our model works for both RGB and RGBD data, and dynamically estimating depth if
not available. Taking the inner workings of depth-prediction networks into
account, we propose to estimate the pseudo-geometry maps from RGB input -
essentially mimicking the multi-modality input. 2) Our MGSnet for RGB SOD
results in real-time inference but achieves state-of-the-art performance
compared to other RGB models. 3) The flexible and lightweight design of MGS
facilitates the integration into RGBD two-streaming models. The introduced
fusion design enables a cross-modality interaction to enable further progress
but with a minimal cost.
Related papers
- MambaSOD: Dual Mamba-Driven Cross-Modal Fusion Network for RGB-D Salient Object Detection [10.848413136031025]
The purpose of RGB-D Salient Object Detection (SOD) is to pinpoint the most visually conspicuous areas within images accurately.
We propose a dual Mamba-driven cross-modal fusion network for RGB-D SOD, named MambaSOD.
arXiv Detail & Related papers (2024-10-19T07:08:40Z) - HODINet: High-Order Discrepant Interaction Network for RGB-D Salient
Object Detection [4.007827908611563]
RGB-D salient object detection (SOD) aims to detect the prominent regions by jointly modeling RGB and depth information.
Most RGB-D SOD methods apply the same type of backbones and fusion modules to identically learn the multimodality and multistage features.
In this paper, we propose a high-order discrepant interaction network (HODINet) for RGB-D SOD.
arXiv Detail & Related papers (2023-07-03T11:56:21Z) - Interactive Context-Aware Network for RGB-T Salient Object Detection [7.544240329265388]
We propose a novel network called Interactive Context-Aware Network (ICANet)
ICANet contains three modules that can effectively perform the cross-modal and cross-scale fusions.
Experiments prove that our network performs favorably against the state-of-the-art RGB-T SOD methods.
arXiv Detail & Related papers (2022-11-11T10:04:36Z) - Dual Swin-Transformer based Mutual Interactive Network for RGB-D Salient
Object Detection [67.33924278729903]
In this work, we propose Dual Swin-Transformer based Mutual Interactive Network.
We adopt Swin-Transformer as the feature extractor for both RGB and depth modality to model the long-range dependencies in visual inputs.
Comprehensive experiments on five standard RGB-D SOD benchmark datasets demonstrate the superiority of the proposed DTMINet method.
arXiv Detail & Related papers (2022-06-07T08:35:41Z) - Cross-modality Discrepant Interaction Network for RGB-D Salient Object
Detection [78.47767202232298]
We propose a novel Cross-modality Discrepant Interaction Network (CDINet) for RGB-D SOD.
Two components are designed to implement the effective cross-modality interaction.
Our network outperforms $15$ state-of-the-art methods both quantitatively and qualitatively.
arXiv Detail & Related papers (2021-08-04T11:24:42Z) - Self-Supervised Representation Learning for RGB-D Salient Object
Detection [93.17479956795862]
We use Self-Supervised Representation Learning to design two pretext tasks: the cross-modal auto-encoder and the depth-contour estimation.
Our pretext tasks require only a few and un RGB-D datasets to perform pre-training, which make the network capture rich semantic contexts.
For the inherent problem of cross-modal fusion in RGB-D SOD, we propose a multi-path fusion module.
arXiv Detail & Related papers (2021-01-29T09:16:06Z) - Bi-directional Cross-Modality Feature Propagation with
Separation-and-Aggregation Gate for RGB-D Semantic Segmentation [59.94819184452694]
Depth information has proven to be a useful cue in the semantic segmentation of RGBD images for providing a geometric counterpart to the RGB representation.
Most existing works simply assume that depth measurements are accurate and well-aligned with the RGB pixels and models the problem as a cross-modal feature fusion.
In this paper, we propose a unified and efficient Crossmodality Guided to not only effectively recalibrate RGB feature responses, but also to distill accurate depth information via multiple stages and aggregate the two recalibrated representations alternatively.
arXiv Detail & Related papers (2020-07-17T18:35:24Z) - Cross-Modal Weighting Network for RGB-D Salient Object Detection [76.0965123893641]
We propose a novel Cross-Modal Weighting (CMW) strategy to encourage comprehensive interactions between RGB and depth channels for RGB-D SOD.
Specifically, three RGB-depth interaction modules, named CMW-L, CMW-M and CMW-H, are developed to deal with respectively low-, middle- and high-level cross-modal information fusion.
CMWNet consistently outperforms 15 state-of-the-art RGB-D SOD methods on seven popular benchmarks.
arXiv Detail & Related papers (2020-07-09T16:01:44Z) - Is Depth Really Necessary for Salient Object Detection? [50.10888549190576]
We make the first attempt in realizing an unified depth-aware framework with only RGB information as input for inference.
Not only surpasses the state-of-the-art performances on five public RGB SOD benchmarks, but also surpasses the RGBD-based methods on five benchmarks by a large margin.
arXiv Detail & Related papers (2020-05-30T13:40:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.