An End-to-End Network for Co-Saliency Detection in One Single Image
- URL: http://arxiv.org/abs/1910.11819v2
- Date: Wed, 15 Feb 2023 15:17:28 GMT
- Title: An End-to-End Network for Co-Saliency Detection in One Single Image
- Authors: Yuanhao Yue, Qin Zou, Hongkai Yu, Qian Wang, Zhongyuan Wang and Song
Wang
- Abstract summary: Co-saliency detection within a single image is a common vision problem that has not yet been well addressed.
This study proposes a novel end-to-end trainable network comprising a backbone net and two branch nets.
We construct a new dataset of 2,019 natural images with co-saliency in each image to evaluate the proposed method.
- Score: 47.35448093528382
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Co-saliency detection within a single image is a common vision problem that
has received little attention and has not yet been well addressed. Existing
methods often used a bottom-up strategy to infer co-saliency in an image in
which salient regions are firstly detected using visual primitives such as
color and shape and then grouped and merged into a co-saliency map. However,
co-saliency is intrinsically perceived complexly with bottom-up and top-down
strategies combined in human vision. To address this problem, this study
proposes a novel end-to-end trainable network comprising a backbone net and two
branch nets. The backbone net uses ground-truth masks as top-down guidance for
saliency prediction, whereas the two branch nets construct triplet proposals
for regional feature mapping and clustering, which drives the network to be
bottom-up sensitive to co-salient regions. We construct a new dataset of 2,019
natural images with co-saliency in each image to evaluate the proposed method.
Experimental results show that the proposed method achieves state-of-the-art
accuracy with a running speed of 28 fps.
Related papers
- Dense Siamese Network [86.23741104851383]
We present Dense Siamese Network (DenseSiam), a simple unsupervised learning framework for dense prediction tasks.
It learns visual representations by maximizing the similarity between two views of one image with two types of consistency, i.e., pixel consistency and region consistency.
It surpasses state-of-the-art segmentation methods by 2.1 mIoU with 28% training costs.
arXiv Detail & Related papers (2022-03-21T15:55:23Z) - Sharp-GAN: Sharpness Loss Regularized GAN for Histopathology Image
Synthesis [65.47507533905188]
Conditional generative adversarial networks have been applied to generate synthetic histopathology images.
We propose a sharpness loss regularized generative adversarial network to synthesize realistic histopathology images.
arXiv Detail & Related papers (2021-10-27T18:54:25Z) - ADMM-DAD net: a deep unfolding network for analysis compressed sensing [20.88999913266683]
We propose a new deep unfolding neural network based on the ADMM algorithm for analysis Compressed Sensing.
The proposed network jointly learns a redundant analysis operator for sparsification and reconstructs the signal of interest.
arXiv Detail & Related papers (2021-10-13T18:56:59Z) - SOSD-Net: Joint Semantic Object Segmentation and Depth Estimation from
Monocular images [94.36401543589523]
We introduce the concept of semantic objectness to exploit the geometric relationship of these two tasks.
We then propose a Semantic Object and Depth Estimation Network (SOSD-Net) based on the objectness assumption.
To the best of our knowledge, SOSD-Net is the first network that exploits the geometry constraint for simultaneous monocular depth estimation and semantic segmentation.
arXiv Detail & Related papers (2021-01-19T02:41:03Z) - A Lightweight Neural Network for Monocular View Generation with
Occlusion Handling [46.74874316127603]
We present a very lightweight neural network architecture, trained on stereo data pairs, which performs view synthesis from one single image.
The work outperforms visually and metric-wise state-of-the-art approaches on the challenging KITTI dataset.
arXiv Detail & Related papers (2020-07-24T15:29:01Z) - SharinGAN: Combining Synthetic and Real Data for Unsupervised Geometry
Estimation [18.29202999419042]
We propose a novel method for combining synthetic and real images when training networks.
We suggest a method for mapping both image types into a single, shared domain.
Our experiments demonstrate significant improvements over the state-of-the-art in two important domains.
arXiv Detail & Related papers (2020-06-07T02:45:33Z) - Ventral-Dorsal Neural Networks: Object Detection via Selective Attention [51.79577908317031]
We propose a new framework called Ventral-Dorsal Networks (VDNets)
Inspired by the structure of the human visual system, we propose the integration of a "Ventral Network" and a "Dorsal Network"
Our experimental results reveal that the proposed method outperforms state-of-the-art object detection approaches.
arXiv Detail & Related papers (2020-05-15T23:57:36Z) - Gradient-Induced Co-Saliency Detection [81.54194063218216]
Co-saliency detection (Co-SOD) aims to segment the common salient foreground in a group of relevant images.
In this paper, inspired by human behavior, we propose a gradient-induced co-saliency detection method.
arXiv Detail & Related papers (2020-04-28T08:40:55Z) - A Two-Stream Symmetric Network with Bidirectional Ensemble for Aerial
Image Matching [24.089374888914143]
We propose a novel method to precisely match two aerial images that were obtained in different environments via a two-stream deep network.
By internally augmenting the target image, the network considers the two-stream with the three input images and reflects the additional augmented pair in the training.
arXiv Detail & Related papers (2020-02-04T14:38:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.