Contrast-weighted Dictionary Learning Based Saliency Detection for
Remote Sensing Images
- URL: http://arxiv.org/abs/2004.02428v2
- Date: Sun, 10 May 2020 07:10:02 GMT
- Title: Contrast-weighted Dictionary Learning Based Saliency Detection for
Remote Sensing Images
- Authors: Zhou Huang, Huai-Xin Chen, Tao Zhou, Yun-Zhi Yang, Chang-Yin Wang and
Bi-Yuan Liu
- Abstract summary: We propose a novel saliency detection model based on Contrast-weighted Dictionary Learning (CDL) for remote sensing images.
Specifically, the proposed CDL learns salient and non-salient atoms from positive and negative samples to construct a discriminant dictionary.
By using the proposed joint saliency measure, a variety of saliency maps are generated based on the discriminant dictionary.
- Score: 3.338193485961624
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Object detection is an important task in remote sensing image analysis. To
reduce the computational complexity of redundant information and improve the
efficiency of image processing, visual saliency models have been widely applied
in this field. In this paper, a novel saliency detection model based on
Contrast-weighted Dictionary Learning (CDL) is proposed for remote sensing
images. Specifically, the proposed CDL learns salient and non-salient atoms
from positive and negative samples to construct a discriminant dictionary, in
which a contrast-weighted term is proposed to encourage the contrast-weighted
patterns to be present in the learned salient dictionary while discouraging
them from being present in the non-salient dictionary. Then, we measure the
saliency by combining the coefficients of the sparse representation (SR) and
reconstruction errors. Furthermore, by using the proposed joint saliency
measure, a variety of saliency maps are generated based on the discriminant
dictionary. Finally, a fusion method based on global gradient optimization is
proposed to integrate multiple saliency maps. Experimental results on four
datasets demonstrate that the proposed model outperforms other state-of-the-art
methods.
Related papers
- Transformer-based Clipped Contrastive Quantization Learning for
Unsupervised Image Retrieval [15.982022297570108]
Unsupervised image retrieval aims to learn the important visual characteristics without any given level to retrieve the similar images for a given query image.
In this paper, we propose a TransClippedCLR model by encoding the global context of an image using Transformer having local context through patch based processing.
Results using the proposed clipped contrastive learning are greatly improved on all datasets as compared to same backbone network with vanilla contrastive learning.
arXiv Detail & Related papers (2024-01-27T09:39:11Z) - A variational autoencoder-based nonnegative matrix factorisation model
for deep dictionary learning [13.796655751448288]
Construction of dictionaries using nonnegative matrix factorisation (NMF) has extensive applications in signal processing and machine learning.
We propose a probabilistic generative model which employs a variational autoencoder (VAE) to perform nonnegative dictionary learning.
arXiv Detail & Related papers (2023-01-18T02:36:03Z) - Deep Equilibrium Assisted Block Sparse Coding of Inter-dependent
Signals: Application to Hyperspectral Imaging [71.57324258813675]
A dataset of inter-dependent signals is defined as a matrix whose columns demonstrate strong dependencies.
A neural network is employed to act as structure prior and reveal the underlying signal interdependencies.
Deep unrolling and Deep equilibrium based algorithms are developed, forming highly interpretable and concise deep-learning-based architectures.
arXiv Detail & Related papers (2022-03-29T21:00:39Z) - Dictionary learning for clustering on hyperspectral images [0.5584060970507506]
We propose a method for clustering the pixels of hyperspectral images using sparse coefficients computed from a representative dictionary as features.
We show empirically that the proposed method works more effectively than clustering on the original pixels.
arXiv Detail & Related papers (2022-02-02T12:22:33Z) - Dictionary Learning with Uniform Sparse Representations for Anomaly
Detection [2.277447144331876]
We study how dictionary learning (DL) performs in detecting abnormal samples in a dataset of signals.
Numerical simulations show that one can efficiently use this resulted subspace to discriminate the anomalies over the regular data points.
arXiv Detail & Related papers (2022-01-11T10:22:46Z) - Bayesian Graph Contrastive Learning [55.36652660268726]
We propose a novel perspective of graph contrastive learning methods showing random augmentations leads to encoders.
Our proposed method represents each node by a distribution in the latent space in contrast to existing techniques which embed each node to a deterministic vector.
We show a considerable improvement in performance compared to existing state-of-the-art methods on several benchmark datasets.
arXiv Detail & Related papers (2021-12-15T01:45:32Z) - Two-stage Visual Cues Enhancement Network for Referring Image
Segmentation [89.49412325699537]
Referring Image (RIS) aims at segmenting the target object from an image referred by one given natural language expression.
In this paper, we tackle this problem by devising a Two-stage Visual cues enhancement Network (TV-Net)
Through the two-stage enhancement, our proposed TV-Net enjoys better performances in learning fine-grained matching behaviors between the natural language expression and image.
arXiv Detail & Related papers (2021-10-09T02:53:39Z) - Deep learning based dictionary learning and tomographic image
reconstruction [0.0]
This work presents an approach for image reconstruction in clinical low-dose tomography that combines principles from sparse signal processing with ideas from deep learning.
First, we describe sparse signal representation in terms of dictionaries from a statistical perspective and interpret dictionary learning as a process of aligning distribution that arises from a generative model with empirical distribution of true signals.
As a result we can see that sparse coding with learned dictionaries resembles a specific variational autoencoder, where the decoder is a linear function and the encoder is a sparse coding algorithm.
arXiv Detail & Related papers (2021-08-26T12:10:17Z) - Object-aware Contrastive Learning for Debiased Scene Representation [74.30741492814327]
We develop a novel object-aware contrastive learning framework that localizes objects in a self-supervised manner.
We also introduce two data augmentations based on ContraCAM, object-aware random crop and background mixup, which reduce contextual and background biases during contrastive self-supervised learning.
arXiv Detail & Related papers (2021-07-30T19:24:07Z) - Blind Face Restoration via Deep Multi-scale Component Dictionaries [75.02640809505277]
We propose a deep face dictionary network (termed as DFDNet) to guide the restoration process of degraded observations.
DFDNet generates deep dictionaries for perceptually significant face components from high-quality images.
component AdaIN is leveraged to eliminate the style diversity between the input and dictionary features.
arXiv Detail & Related papers (2020-08-02T07:02:07Z) - Distilling Localization for Self-Supervised Representation Learning [82.79808902674282]
Contrastive learning has revolutionized unsupervised representation learning.
Current contrastive models are ineffective at localizing the foreground object.
We propose a data-driven approach for learning in variance to backgrounds.
arXiv Detail & Related papers (2020-04-14T16:29:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.