Credible Remote Sensing Scene Classification Using Evidential Fusion on
Aerial-Ground Dual-view Images
- URL: http://arxiv.org/abs/2301.00622v1
- Date: Mon, 2 Jan 2023 12:27:55 GMT
- Title: Credible Remote Sensing Scene Classification Using Evidential Fusion on
Aerial-Ground Dual-view Images
- Authors: Kun Zhao, Qian Gao, Siyuan Hao, Jie Sun, Lijian Zhou
- Abstract summary: Multi-view (multi-source, multi-modal, multi-perspective, etc.) data are being used more frequently in remote sensing tasks.
The issue of data quality becomes more apparent, limiting the potential benefits of multi-view data.
Deep learning is introduced to the task of aerial-ground dual-view remote sensing scene classification.
- Score: 6.817740582240199
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Due to their ability to offer more comprehensive information than data from a
single view, multi-view (multi-source, multi-modal, multi-perspective, etc.)
data are being used more frequently in remote sensing tasks. However, as the
number of views grows, the issue of data quality becomes more apparent,
limiting the potential benefits of multi-view data. Although recent deep neural
network (DNN) based models can learn the weight of data adaptively, a lack of
research on explicitly quantifying the data quality of each view when fusing
them renders these models inexplicable, performing unsatisfactorily and
inflexible in downstream remote sensing tasks. To fill this gap, in this paper,
evidential deep learning is introduced to the task of aerial-ground dual-view
remote sensing scene classification to model the credibility of each view.
Specifically, the theory of evidence is used to calculate an uncertainty value
which describes the decision-making risk of each view. Based on this
uncertainty, a novel decision-level fusion strategy is proposed to ensure that
the view with lower risk obtains more weight, making the classification more
credible. On two well-known, publicly available datasets of aerial-ground
dual-view remote sensing images, the proposed approach achieves
state-of-the-art results, demonstrating its effectiveness. The code and
datasets of this article are available at the following address:
https://github.com/gaopiaoliang/Evidential.
Related papers
- SpecRaGE: Robust and Generalizable Multi-view Spectral Representation Learning [9.393841121141076]
Multi-view representation learning (MvRL) has garnered substantial attention in recent years.
graph Laplacian-based MvRL methods have demonstrated remarkable success in representing multi-view data.
We introduce $textitSpecRaGE$, a novel fusion-based framework that integrates the strengths of graph Laplacian methods with the power of deep learning.
arXiv Detail & Related papers (2024-11-04T14:51:35Z) - Regularized Contrastive Partial Multi-view Outlier Detection [76.77036536484114]
We propose a novel method named Regularized Contrastive Partial Multi-view Outlier Detection (RCPMOD)
In this framework, we utilize contrastive learning to learn view-consistent information and distinguish outliers by the degree of consistency.
Experimental results on four benchmark datasets demonstrate that our proposed approach could outperform state-of-the-art competitors.
arXiv Detail & Related papers (2024-08-02T14:34:27Z) - Towards General Visual-Linguistic Face Forgery Detection [95.73987327101143]
Deepfakes are realistic face manipulations that can pose serious threats to security, privacy, and trust.
Existing methods mostly treat this task as binary classification, which uses digital labels or mask signals to train the detection model.
We propose a novel paradigm named Visual-Linguistic Face Forgery Detection(VLFFD), which uses fine-grained sentence-level prompts as the annotation.
arXiv Detail & Related papers (2023-07-31T10:22:33Z) - Two-level Data Augmentation for Calibrated Multi-view Detection [51.5746691103591]
We introduce a new multi-view data augmentation pipeline that preserves alignment among views.
We also propose a second level of augmentation applied directly at the scene level.
When combined with our simple multi-view detection model, our two-level augmentation pipeline outperforms all existing baselines.
arXiv Detail & Related papers (2022-10-19T17:55:13Z) - Mitigating Representation Bias in Action Recognition: Algorithms and
Benchmarks [76.35271072704384]
Deep learning models perform poorly when applied to videos with rare scenes or objects.
We tackle this problem from two different angles: algorithm and dataset.
We show that the debiased representation can generalize better when transferred to other datasets and tasks.
arXiv Detail & Related papers (2022-09-20T00:30:35Z) - Facing the Void: Overcoming Missing Data in Multi-View Imagery [0.783788180051711]
We propose a novel technique for multi-view image classification robust to this problem.
The proposed method, based on state-of-the-art deep learning-based approaches and metric learning, can be easily adapted and exploited in other applications and domains.
Results show that the proposed algorithm provides improvements in multi-view image classification accuracy when compared to state-of-the-art methods.
arXiv Detail & Related papers (2022-05-21T13:21:27Z) - Uncertainty-Aware Multi-View Representation Learning [53.06828186507994]
We devise a novel unsupervised multi-view learning approach, termed as Dynamic Uncertainty-Aware Networks (DUA-Nets)
Guided by the uncertainty of data estimated from the generation perspective, intrinsic information from multiple views is integrated to obtain noise-free representations.
Our model achieves superior performance in extensive experiments and shows the robustness to noisy data.
arXiv Detail & Related papers (2022-01-15T07:16:20Z) - Generative Partial Multi-View Clustering [133.36721417531734]
We propose a generative partial multi-view clustering model, named as GP-MVC, to address the incomplete multi-view problem.
First, multi-view encoder networks are trained to learn common low-dimensional representations, followed by a clustering layer to capture the consistent cluster structure across multiple views.
Second, view-specific generative adversarial networks are developed to generate the missing data of one view conditioning on the shared representation given by other views.
arXiv Detail & Related papers (2020-03-29T17:48:27Z) - Learning Robust Representations via Multi-View Information Bottleneck [41.65544605954621]
Original formulation requires labeled data to identify superfluous information.
We extend this ability to the multi-view unsupervised setting, where two views of the same underlying entity are provided but the label is unknown.
A theoretical analysis leads to the definition of a new multi-view model that produces state-of-the-art results on the Sketchy dataset and label-limited versions of the MIR-Flickr dataset.
arXiv Detail & Related papers (2020-02-17T16:01:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.