Camouflaged Object Detection via Context-aware Cross-level Fusion
- URL: http://arxiv.org/abs/2207.13362v1
- Date: Wed, 27 Jul 2022 08:34:16 GMT
- Title: Camouflaged Object Detection via Context-aware Cross-level Fusion
- Authors: Geng Chen, Si-Jie Liu, Yu-Jia Sun, Ge-Peng Ji, Ya-Feng Wu, Tao Zhou
- Abstract summary: Camouflaged object detection (COD) aims to identify the objects that conceal themselves in natural scenes.
We propose a novel Context-aware Cross-level Fusion Network (C2F-Net), which fuses context-aware cross-level features.
C2F-Net is an effective COD model and outperforms state-of-the-art (SOTA) models remarkably.
- Score: 10.942917945534678
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Camouflaged object detection (COD) aims to identify the objects that conceal
themselves in natural scenes. Accurate COD suffers from a number of challenges
associated with low boundary contrast and the large variation of object
appearances, e.g., object size and shape. To address these challenges, we
propose a novel Context-aware Cross-level Fusion Network (C2F-Net), which fuses
context-aware cross-level features for accurately identifying camouflaged
objects. Specifically, we compute informative attention coefficients from
multi-level features with our Attention-induced Cross-level Fusion Module
(ACFM), which further integrates the features under the guidance of attention
coefficients. We then propose a Dual-branch Global Context Module (DGCM) to
refine the fused features for informative feature representations by exploiting
rich global context information. Multiple ACFMs and DGCMs are integrated in a
cascaded manner for generating a coarse prediction from high-level features.
The coarse prediction acts as an attention map to refine the low-level features
before passing them to our Camouflage Inference Module (CIM) to generate the
final prediction. We perform extensive experiments on three widely used
benchmark datasets and compare C2F-Net with state-of-the-art (SOTA) models. The
results show that C2F-Net is an effective COD model and outperforms SOTA models
remarkably. Further, an evaluation on polyp segmentation datasets demonstrates
the promising potentials of our C2F-Net in COD downstream applications. Our
code is publicly available at: https://github.com/Ben57882/C2FNet-TSCVT.
Related papers
- PVAFN: Point-Voxel Attention Fusion Network with Multi-Pooling Enhancing for 3D Object Detection [59.355022416218624]
integration of point and voxel representations is becoming more common in LiDAR-based 3D object detection.
We propose a novel two-stage 3D object detector, called Point-Voxel Attention Fusion Network (PVAFN)
PVAFN uses a multi-pooling strategy to integrate both multi-scale and region-specific information effectively.
arXiv Detail & Related papers (2024-08-26T19:43:01Z) - CoFiNet: Unveiling Camouflaged Objects with Multi-Scale Finesse [46.79770062391987]
We introduce a novel method for camouflage object detection, named CoFiNet.
Our approach focuses on multi-scale feature fusion and extraction, with special attention to the model's segmentation effectiveness.
CoFiNet achieves state-of-the-art performance across all datasets.
arXiv Detail & Related papers (2024-02-03T17:24:55Z) - Salient Object Detection in Optical Remote Sensing Images Driven by
Transformer [69.22039680783124]
We propose a novel Global Extraction Local Exploration Network (GeleNet) for Optical Remote Sensing Images (ORSI-SOD)
Specifically, GeleNet first adopts a transformer backbone to generate four-level feature embeddings with global long-range dependencies.
Extensive experiments on three public datasets demonstrate that the proposed GeleNet outperforms relevant state-of-the-art methods.
arXiv Detail & Related papers (2023-09-15T07:14:43Z) - Frequency Perception Network for Camouflaged Object Detection [51.26386921922031]
We propose a novel learnable and separable frequency perception mechanism driven by the semantic hierarchy in the frequency domain.
Our entire network adopts a two-stage model, including a frequency-guided coarse localization stage and a detail-preserving fine localization stage.
Compared with the currently existing models, our proposed method achieves competitive performance in three popular benchmark datasets.
arXiv Detail & Related papers (2023-08-17T11:30:46Z) - Feature Aggregation and Propagation Network for Camouflaged Object
Detection [42.33180748293329]
Camouflaged object detection (COD) aims to detect/segment camouflaged objects embedded in the environment.
Several COD methods have been developed, but they still suffer from unsatisfactory performance due to intrinsic similarities between foreground objects and background surroundings.
We propose a novel Feature Aggregation and propagation Network (FAP-Net) for camouflaged object detection.
arXiv Detail & Related papers (2022-12-02T05:54:28Z) - EPNet++: Cascade Bi-directional Fusion for Multi-Modal 3D Object
Detection [56.03081616213012]
We propose EPNet++ for multi-modal 3D object detection by introducing a novel Cascade Bi-directional Fusion(CB-Fusion) module.
The proposed CB-Fusion module boosts the plentiful semantic information of point features with the image features in a cascade bi-directional interaction fusion manner.
The experiment results on the KITTI, JRDB and SUN-RGBD datasets demonstrate the superiority of EPNet++ over the state-of-the-art methods.
arXiv Detail & Related papers (2021-12-21T10:48:34Z) - Context-aware Cross-level Fusion Network for Camouflaged Object
Detection [23.109969322128958]
We propose a novel Context-aware Cross-level Fusion Network (C2F-Net) to address the challenging COD task.
We show that our C2F-Net is an effective COD model and outperforms state-of-the-art models remarkably.
arXiv Detail & Related papers (2021-05-26T14:03:36Z) - Towards Accurate Camouflaged Object Detection with Mixture Convolution and Interactive Fusion [45.45231015502287]
We propose a novel deep learning based COD approach, which integrates the large receptive field and effective feature fusion into a unified framework.
Our method detects camouflaged objects with an effective fusion strategy, which aggregates the rich context information from a large receptive field.
arXiv Detail & Related papers (2021-01-14T16:06:08Z) - Learning Selective Mutual Attention and Contrast for RGB-D Saliency
Detection [145.4919781325014]
How to effectively fuse cross-modal information is the key problem for RGB-D salient object detection.
Many models use the feature fusion strategy but are limited by the low-order point-to-point fusion methods.
We propose a novel mutual attention model by fusing attention and contexts from different modalities.
arXiv Detail & Related papers (2020-10-12T08:50:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.