Camouflaged Object Segmentation with Distraction Mining
- URL: http://arxiv.org/abs/2104.10475v1
- Date: Wed, 21 Apr 2021 11:47:59 GMT
- Title: Camouflaged Object Segmentation with Distraction Mining
- Authors: Haiyang Mei, Ge-Peng Ji, Ziqi Wei, Xin Yang, Xiaopeng Wei, Deng-Ping
Fan
- Abstract summary: Camouflaged object segmentation (COS) aims to identify objects that are "perfectly" assimilate into their surroundings.
We develop a bio-inspired framework, termed Positioning and Focus Network (PFNet), which mimics the process of predation in nature.
Our PFNet runs in real-time (72 FPS) and significantly outperforms 18 cutting-edge models on three challenging datasets.
- Score: 23.77915054363188
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Camouflaged object segmentation (COS) aims to identify objects that are
"perfectly" assimilate into their surroundings, which has a wide range of
valuable applications. The key challenge of COS is that there exist high
intrinsic similarities between the candidate objects and noise background. In
this paper, we strive to embrace challenges towards effective and efficient
COS. To this end, we develop a bio-inspired framework, termed Positioning and
Focus Network (PFNet), which mimics the process of predation in nature.
Specifically, our PFNet contains two key modules, i.e., the positioning module
(PM) and the focus module (FM). The PM is designed to mimic the detection
process in predation for positioning the potential target objects from a global
perspective and the FM is then used to perform the identification process in
predation for progressively refining the coarse prediction via focusing on the
ambiguous regions. Notably, in the FM, we develop a novel distraction mining
strategy for distraction discovery and removal, to benefit the performance of
estimation. Extensive experiments demonstrate that our PFNet runs in real-time
(72 FPS) and significantly outperforms 18 cutting-edge models on three
challenging datasets under four standard metrics.
Related papers
- 3rd Place Solution for MOSE Track in CVPR 2024 PVUW workshop: Complex Video Object Segmentation [63.199793919573295]
Video Object (VOS) is a vital task in computer vision, focusing on distinguishing foreground objects from the background across video frames.
Our work draws inspiration from the Cutie model, and we investigate the effects of object memory, the total number of memory frames, and input resolution on segmentation performance.
arXiv Detail & Related papers (2024-06-06T00:56:25Z) - Large receptive field strategy and important feature extraction strategy
in 3D object detection [6.3948571459793975]
This study focuses on key challenges in 3D target detection.
To tackle the challenge of expanding the receptive field of a 3D convolutional kernel, we introduce the Dynamic Feature Fusion Module.
This module achieves adaptive expansion of the 3D convolutional kernel's receptive field, balancing the expansion with acceptable computational loads.
arXiv Detail & Related papers (2024-01-22T13:01:28Z) - Frequency Perception Network for Camouflaged Object Detection [51.26386921922031]
We propose a novel learnable and separable frequency perception mechanism driven by the semantic hierarchy in the frequency domain.
Our entire network adopts a two-stage model, including a frequency-guided coarse localization stage and a detail-preserving fine localization stage.
Compared with the currently existing models, our proposed method achieves competitive performance in three popular benchmark datasets.
arXiv Detail & Related papers (2023-08-17T11:30:46Z) - Object Preserving Siamese Network for Single Object Tracking on Point
Clouds [0.6165605009782557]
We propose an Object Preserving Siamese Network (OPSNet), which can significantly maintain object integrity and boost tracking performance.
First, the object highlighting module enhances the object-aware features and extracts discriminative features from template and search area.
Then, the object-preserved sampling selects object candidates to obtain object-preserved search area seeds and drop the background points that contribute less to tracking.
Finally, the object localization network precisely locates 3D BBoxes based on the object-preserved search area seeds.
arXiv Detail & Related papers (2023-01-28T02:21:31Z) - A Tri-Layer Plugin to Improve Occluded Detection [100.99802831241583]
We propose a simple '' module for the detection head of two-stage object detectors to improve the recall of partially occluded objects.
The module predicts a tri-layer of segmentation masks for the target object, the occluder and the occludee, and by doing so is able to better predict the mask of the target object.
We also establish a COCO evaluation dataset to measure the recall performance of partially occluded and separated objects.
arXiv Detail & Related papers (2022-10-18T17:59:51Z) - Centralized Feature Pyramid for Object Detection [53.501796194901964]
Visual feature pyramid has shown its superiority in both effectiveness and efficiency in a wide range of applications.
In this paper, we propose a OLO Feature Pyramid for object detection, which is based on a globally explicit centralized feature regulation.
arXiv Detail & Related papers (2022-10-05T08:32:54Z) - Oriented Feature Alignment for Fine-grained Object Recognition in
High-Resolution Satellite Imagery [1.0635248457021498]
We analyze the key issues of fine-grained object recognition, and use an oriented feature alignment network (OFA-Net) to achieve high-performance object recognition.
OFA-Net achieves accurate object localization through a rotated bounding boxes refinement module.
The single model of our method achieved mAP of 46.51% in the GaoFen competition and won 3rd place in the ISPRS benchmark with the mAP of 43.73%.
arXiv Detail & Related papers (2021-10-13T10:48:11Z) - PC-RGNN: Point Cloud Completion and Graph Neural Network for 3D Object
Detection [57.49788100647103]
LiDAR-based 3D object detection is an important task for autonomous driving.
Current approaches suffer from sparse and partial point clouds of distant and occluded objects.
In this paper, we propose a novel two-stage approach, namely PC-RGNN, dealing with such challenges by two specific solutions.
arXiv Detail & Related papers (2020-12-18T18:06:43Z) - End-to-End Object Detection with Fully Convolutional Network [71.56728221604158]
We introduce a Prediction-aware One-To-One (POTO) label assignment for classification to enable end-to-end detection.
A simple 3D Max Filtering (3DMF) is proposed to utilize the multi-scale features and improve the discriminability of convolutions in the local region.
Our end-to-end framework achieves competitive performance against many state-of-the-art detectors with NMS on COCO and CrowdHuman datasets.
arXiv Detail & Related papers (2020-12-07T09:14:55Z) - Attention-based Assisted Excitation for Salient Object Detection [3.238929552408813]
We introduce a mechanism for modification of activations in feature maps of CNNs inspired by object-based attention in brain.
Similar to brain, we use the idea to address two challenges in salient object detection: gathering object interior parts while segregation from background with concise boundaries.
We implement the object-based attention in the U-net model using different architectures in the encoder parts, including AlexNet, VGG, and ResNet.
arXiv Detail & Related papers (2020-03-31T13:33:33Z) - PENet: Object Detection using Points Estimation in Aerial Images [9.33900415971554]
A novel network structure, Points Estimated Network (PENet), is proposed in this work to answer these challenges.
PENet uses a Mask Resampling Module (MRM) to augment the imbalanced datasets, a coarse anchor-free detector (CPEN) to effectively predict the center points of the small object clusters, and a fine anchor-free detector FPEN to locate the precise positions of the small objects.
Our experiments on aerial datasets visDrone and UAVDT showed that PENet achieved higher precision results than existing state-of-the-art approaches.
arXiv Detail & Related papers (2020-01-22T19:43:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.