Uncertainty Inspired RGB-D Saliency Detection
- URL: http://arxiv.org/abs/2009.03075v1
- Date: Mon, 7 Sep 2020 13:01:45 GMT
- Title: Uncertainty Inspired RGB-D Saliency Detection
- Authors: Jing Zhang, Deng-Ping Fan, Yuchao Dai, Saeed Anwar, Fatemeh Saleh,
Sadegh Aliakbarian, Nick Barnes
- Abstract summary: We propose the first framework to employ uncertainty for RGB-D saliency detection by learning from the data labeling process.
Inspired by the saliency data labeling process, we propose a generative architecture to achieve probabilistic RGB-D saliency detection.
Results on six challenging RGB-D benchmark datasets show our approach's superior performance in learning the distribution of saliency maps.
- Score: 70.50583438784571
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose the first stochastic framework to employ uncertainty for RGB-D
saliency detection by learning from the data labeling process. Existing RGB-D
saliency detection models treat this task as a point estimation problem by
predicting a single saliency map following a deterministic learning pipeline.
We argue that, however, the deterministic solution is relatively ill-posed.
Inspired by the saliency data labeling process, we propose a generative
architecture to achieve probabilistic RGB-D saliency detection which utilizes a
latent variable to model the labeling variations. Our framework includes two
main models: 1) a generator model, which maps the input image and latent
variable to stochastic saliency prediction, and 2) an inference model, which
gradually updates the latent variable by sampling it from the true or
approximate posterior distribution. The generator model is an encoder-decoder
saliency network. To infer the latent variable, we introduce two different
solutions: i) a Conditional Variational Auto-encoder with an extra encoder to
approximate the posterior distribution of the latent variable; and ii) an
Alternating Back-Propagation technique, which directly samples the latent
variable from the true posterior distribution. Qualitative and quantitative
results on six challenging RGB-D benchmark datasets show our approach's
superior performance in learning the distribution of saliency maps. The source
code is publicly available via our project page:
https://github.com/JingZhang617/UCNet.
Related papers
- Tilt your Head: Activating the Hidden Spatial-Invariance of Classifiers [0.7704032792820767]
Deep neural networks are applied in more and more areas of everyday life.
They still lack essential abilities, such as robustly dealing with spatially transformed input signals.
We propose a novel technique to emulate such an inference process for neural nets.
arXiv Detail & Related papers (2024-05-06T09:47:29Z) - An Energy-Based Prior for Generative Saliency [62.79775297611203]
We propose a novel generative saliency prediction framework that adopts an informative energy-based model as a prior distribution.
With the generative saliency model, we can obtain a pixel-wise uncertainty map from an image, indicating model confidence in the saliency prediction.
Experimental results show that our generative saliency model with an energy-based prior can achieve not only accurate saliency predictions but also reliable uncertainty maps consistent with human perception.
arXiv Detail & Related papers (2022-04-19T10:51:00Z) - Performance of Bayesian linear regression in a model with mismatch [8.60118148262922]
We analyze the performance of an estimator given by the mean of a log-concave Bayesian posterior distribution with gaussian prior.
This inference model can be rephrased as a version of the Gardner model in spin glasses.
arXiv Detail & Related papers (2021-07-14T18:50:13Z) - Visual Saliency Transformer [127.33678448761599]
We develop a novel unified model based on a pure transformer, Visual Saliency Transformer (VST), for both RGB and RGB-D salient object detection (SOD)
It takes image patches as inputs and leverages the transformer to propagate global contexts among image patches.
Experimental results show that our model outperforms existing state-of-the-art results on both RGB and RGB-D SOD benchmark datasets.
arXiv Detail & Related papers (2021-04-25T08:24:06Z) - Score-Based Generative Modeling through Stochastic Differential
Equations [114.39209003111723]
We present a differential equation that transforms a complex data distribution to a known prior distribution by injecting noise.
A corresponding reverse-time SDE transforms the prior distribution back into the data distribution by slowly removing the noise.
By leveraging advances in score-based generative modeling, we can accurately estimate these scores with neural networks.
We demonstrate high fidelity generation of 1024 x 1024 images for the first time from a score-based generative model.
arXiv Detail & Related papers (2020-11-26T19:39:10Z) - CQ-VAE: Coordinate Quantized VAE for Uncertainty Estimation with
Application to Disk Shape Analysis from Lumbar Spine MRI Images [1.5841288368322592]
We propose a powerful generative model to learn a representation of ambiguity and to generate probabilistic outputs.
Our model, named Coordinate Quantization Variational Autoencoder (CQ-VAE), employs a discrete latent space with an internal discrete probability distribution.
A matching algorithm is used to establish the correspondence between model-generated samples and "ground-truth" samples.
arXiv Detail & Related papers (2020-10-17T04:25:32Z) - Learning Noise-Aware Encoder-Decoder from Noisy Labels by Alternating
Back-Propagation for Saliency Detection [54.98042023365694]
We propose a noise-aware encoder-decoder framework to disentangle a clean saliency predictor from noisy training examples.
The proposed model consists of two sub-models parameterized by neural networks.
arXiv Detail & Related papers (2020-07-23T18:47:36Z) - UC-Net: Uncertainty Inspired RGB-D Saliency Detection via Conditional
Variational Autoencoders [81.5490760424213]
We propose the first framework (UCNet) to employ uncertainty for RGB-D saliency detection by learning from the data labeling process.
Inspired by the saliency data labeling process, we propose probabilistic RGB-D saliency detection network.
arXiv Detail & Related papers (2020-04-13T04:12:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.