C$^{4}$Net: Contextual Compression and Complementary Combination Network
for Salient Object Detection
- URL: http://arxiv.org/abs/2110.11887v1
- Date: Fri, 22 Oct 2021 16:14:10 GMT
- Title: C$^{4}$Net: Contextual Compression and Complementary Combination Network
for Salient Object Detection
- Authors: Hazarapet Tunanyan
- Abstract summary: We show that feature concatenation works better than other combination methods like multiplication or addition.
Also, joint feature learning gives better results, because of the information sharing during their processing.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning solutions of the salient object detection problem have achieved
great results in recent years. The majority of these models are based on
encoders and decoders, with a different multi-feature combination. In this
paper, we show that feature concatenation works better than other combination
methods like multiplication or addition. Also, joint feature learning gives
better results, because of the information sharing during their processing. We
designed a Complementary Extraction Module (CEM) to extract necessary features
with edge preservation. Our proposed Excessiveness Loss (EL) function helps to
reduce false-positive predictions and purifies the edges with other weighted
loss functions. Our designed Pyramid-Semantic Module (PSM) with Global guiding
flow (G) makes the prediction more accurate by providing high-level
complementary information to shallower layers. Experimental results show that
the proposed model outperforms the state-of-the-art methods on all benchmark
datasets under three evaluation metrics.
Related papers
- RGM: A Robust Generalizable Matching Model [49.60975442871967]
We propose a deep model for sparse and dense matching, termed RGM (Robust Generalist Matching)
To narrow the gap between synthetic training samples and real-world scenarios, we build a new, large-scale dataset with sparse correspondence ground truth.
We are able to mix up various dense and sparse matching datasets, significantly improving the training diversity.
arXiv Detail & Related papers (2023-10-18T07:30:08Z) - Class Anchor Margin Loss for Content-Based Image Retrieval [97.81742911657497]
We propose a novel repeller-attractor loss that falls in the metric learning paradigm, yet directly optimize for the L2 metric without the need of generating pairs.
We evaluate the proposed objective in the context of few-shot and full-set training on the CBIR task, by using both convolutional and transformer architectures.
arXiv Detail & Related papers (2023-06-01T12:53:10Z) - Triplet Loss-less Center Loss Sampling Strategies in Facial Expression
Recognition Scenarios [5.672538282456803]
Deep neural network (DNN) accompanied by deep metric learning (DML) techniques boost the discriminative ability of the model inFER applications.
We developed three strategies: fully-synthesized, semi-synthesized, and prediction-based negative sample selection strategies.
To achieve better results, we introduce a selective attention module that provides a combination of pixel-wise and element-wise attention coefficients.
arXiv Detail & Related papers (2023-02-08T15:03:36Z) - ScoreMix: A Scalable Augmentation Strategy for Training GANs with
Limited Data [93.06336507035486]
Generative Adversarial Networks (GANs) typically suffer from overfitting when limited training data is available.
We present ScoreMix, a novel and scalable data augmentation approach for various image synthesis tasks.
arXiv Detail & Related papers (2022-10-27T02:55:15Z) - Deep Reinforcement Learning of Graph Matching [63.469961545293756]
Graph matching (GM) under node and pairwise constraints has been a building block in areas from optimization to computer vision.
We present a reinforcement learning solver for GM i.e. RGM that seeks the node correspondence between pairwise graphs.
Our method differs from the previous deep graph matching model in the sense that they are focused on the front-end feature extraction and affinity function learning.
arXiv Detail & Related papers (2020-12-16T13:48:48Z) - MAGNeto: An Efficient Deep Learning Method for the Extractive Tags
Summarization Problem [0.0]
We study a new image annotation task named Extractive Tags Summarization (ETS)
The goal is to extract important tags from the context lying in an image and its corresponding tags.
Our proposed solution consists of different widely used blocks like convolutional and self-attention layers.
arXiv Detail & Related papers (2020-11-09T11:34:21Z) - Multi-scale Interactive Network for Salient Object Detection [91.43066633305662]
We propose the aggregate interaction modules to integrate the features from adjacent levels.
To obtain more efficient multi-scale features, the self-interaction modules are embedded in each decoder unit.
Experimental results on five benchmark datasets demonstrate that the proposed method without any post-processing performs favorably against 23 state-of-the-art approaches.
arXiv Detail & Related papers (2020-07-17T15:41:37Z) - Hierarchical Dynamic Filtering Network for RGB-D Salient Object
Detection [91.43066633305662]
The main purpose of RGB-D salient object detection (SOD) is how to better integrate and utilize cross-modal fusion information.
In this paper, we explore these issues from a new perspective.
We implement a kind of more flexible and efficient multi-scale cross-modal feature processing.
arXiv Detail & Related papers (2020-07-13T07:59:55Z) - Multi-Person Pose Estimation with Enhanced Feature Aggregation and
Selection [33.15192824888279]
We propose a novel Enhanced Feature Aggregation and Selection network (EFASNet) for multi-person 2D human pose estimation.
Our method can well handle crowded, cluttered and occluded scenes.
Comprehensive experiments demonstrate that the proposed approach outperforms the state-of-the-art methods.
arXiv Detail & Related papers (2020-03-20T08:33:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.