Co-Saliency Detection with Co-Attention Fully Convolutional Network
- URL: http://arxiv.org/abs/2008.08909v1
- Date: Thu, 20 Aug 2020 11:52:40 GMT
- Title: Co-Saliency Detection with Co-Attention Fully Convolutional Network
- Authors: Guangshuai Gao, Wenting Zhao, Qingjie Liu, Yunhong Wang
- Abstract summary: Co-saliency detection aims to detect common salient objects from a group of relevant images.
We propose a co-attention module embedded FCN framework, called as Co-Attention FCN (CA-FCN)
- Score: 47.26829884104539
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Co-saliency detection aims to detect common salient objects from a group of
relevant images. Some attempts have been made with the Fully Convolutional
Network (FCN) framework and achieve satisfactory detection results. However,
due to stacking convolution layers and pooling operation, the boundary details
tend to be lost. In addition, existing models often utilize the extracted
features without discrimination, leading to redundancy in representation since
actually not all features are helpful to the final prediction and some even
bring distraction. In this paper, we propose a co-attention module embedded FCN
framework, called as Co-Attention FCN (CA-FCN). Specifically, the co-attention
module is plugged into the high-level convolution layers of FCN, which can
assign larger attention weights on the common salient objects and smaller ones
on the background and uncommon distractors to boost final detection
performance. Extensive experiments on three popular co-saliency benchmark
datasets demonstrate the superiority of the proposed CA-FCN, which outperforms
state-of-the-arts in most cases. Besides, the effectiveness of our new
co-attention module is also validated with ablation studies.
Related papers
- CAVE: Classifying Abnormalities in Video Capsule Endoscopy [0.1937002985471497]
In this study, we explore an ensemble-based approach to improve classification accuracy in complex image datasets.
We leverage the unique feature-extraction capabilities of each model to enhance the overall accuracy.
Experimental evaluations demonstrate that the ensemble achieves higher accuracy and robustness across challenging and imbalanced classes.
arXiv Detail & Related papers (2024-10-26T17:25:08Z) - Renormalized Connection for Scale-preferred Object Detection in Satellite Imagery [51.83786195178233]
We design a Knowledge Discovery Network (KDN) to implement the renormalization group theory in terms of efficient feature extraction.
Renormalized connection (RC) on the KDN enables synergistic focusing'' of multi-scale features.
RCs extend the multi-level feature's divide-and-conquer'' mechanism of the FPN-based detectors to a wide range of scale-preferred tasks.
arXiv Detail & Related papers (2024-09-09T13:56:22Z) - AttenScribble: Attentive Similarity Learning for Scribble-Supervised
Medical Image Segmentation [5.8447004333496855]
In this paper, we present a straightforward yet effective scribble supervised learning framework.
We create a pluggable spatial self-attention module which could be attached on top of any internal feature layers of arbitrary fully convolutional network (FCN) backbone.
This attentive similarity leads to a novel regularization loss that imposes consistency between segmentation prediction and visual affinity.
arXiv Detail & Related papers (2023-12-11T18:42:18Z) - GCoNet+: A Stronger Group Collaborative Co-Salient Object Detector [156.43671738038657]
We present a novel end-to-end group collaborative learning network, termed GCoNet+.
GCoNet+ can effectively and efficiently identify co-salient objects in natural scenes.
arXiv Detail & Related papers (2022-05-30T23:49:19Z) - NL-FCOS: Improving FCOS through Non-Local Modules for Object Detection [0.0]
We show that non-local modules combined with an FCOS head (NL-FCOS) are practical and efficient.
We establish state-of-the-art performance in clothing detection and handwritten amount recognition problems.
arXiv Detail & Related papers (2022-03-29T15:00:14Z) - DeFRCN: Decoupled Faster R-CNN for Few-Shot Object Detection [17.326702469604676]
Few-shot object detection aims at detecting novel objects rapidly from extremely few examples of previously unseen classes.
Most existing approaches employ the Faster R-CNN as basic detection framework.
We propose a simple yet effective architecture named Decoupled Faster R-CNN (DeFRCN)
arXiv Detail & Related papers (2021-08-20T06:12:55Z) - Adversarial Feature Augmentation and Normalization for Visual
Recognition [109.6834687220478]
Recent advances in computer vision take advantage of adversarial data augmentation to ameliorate the generalization ability of classification models.
Here, we present an effective and efficient alternative that advocates adversarial augmentation on intermediate feature embeddings.
We validate the proposed approach across diverse visual recognition tasks with representative backbone networks.
arXiv Detail & Related papers (2021-03-22T20:36:34Z) - Towards Accurate RGB-D Saliency Detection with Complementary Attention
and Adaptive Integration [20.006932559837516]
Saliency detection based on the complementary information from RGB images and depth maps has recently gained great popularity.
We propose Complementary Attention and Adaptive Integration Network (CAAI-Net) to integrate complementary attention based feature concentration and adaptive cross-modal feature fusion.
CAAI-Net is an effective saliency detection model and outperforms nine state-of-the-art models in terms of four widely-used metrics.
arXiv Detail & Related papers (2021-02-08T08:08:30Z) - PC-RGNN: Point Cloud Completion and Graph Neural Network for 3D Object
Detection [57.49788100647103]
LiDAR-based 3D object detection is an important task for autonomous driving.
Current approaches suffer from sparse and partial point clouds of distant and occluded objects.
In this paper, we propose a novel two-stage approach, namely PC-RGNN, dealing with such challenges by two specific solutions.
arXiv Detail & Related papers (2020-12-18T18:06:43Z) - End-to-End Object Detection with Fully Convolutional Network [71.56728221604158]
We introduce a Prediction-aware One-To-One (POTO) label assignment for classification to enable end-to-end detection.
A simple 3D Max Filtering (3DMF) is proposed to utilize the multi-scale features and improve the discriminability of convolutions in the local region.
Our end-to-end framework achieves competitive performance against many state-of-the-art detectors with NMS on COCO and CrowdHuman datasets.
arXiv Detail & Related papers (2020-12-07T09:14:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.