Pairwise Comparison Network for Remote Sensing Scene Classification
- URL: http://arxiv.org/abs/2205.08147v1
- Date: Tue, 17 May 2022 07:31:36 GMT
- Title: Pairwise Comparison Network for Remote Sensing Scene Classification
- Authors: Zhang Yue, Zheng Xiangtao, Lu Xiaoqiang
- Abstract summary: This paper proposes a pairwise comparison network, which contains two main steps: pairwise selection and pairwise representation.
The proposed network first selects similar image pairs, and then represents the image pairs with pairwise representations.
The self-representation is introduced to highlight the informative parts of each image itself, while the mutual-representation is proposed to capture the subtle differences between image pairs.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Remote sensing scene classification aims to assign a specific semantic label
to a remote sensing image. Recently, convolutional neural networks have greatly
improved the performance of remote sensing scene classification. However, some
confused images may be easily recognized as the incorrect category, which
generally degrade the performance. The differences between image pairs can be
used to distinguish image categories. This paper proposed a pairwise comparison
network, which contains two main steps: pairwise selection and pairwise
representation. The proposed network first selects similar image pairs, and
then represents the image pairs with pairwise representations. The
self-representation is introduced to highlight the informative parts of each
image itself, while the mutual-representation is proposed to capture the subtle
differences between image pairs. Comprehensive experimental results on two
challenging datasets (AID, NWPU-RESISC45) demonstrate the effectiveness of the
proposed network. The code are provided in
https://github.com/spectralpublic/PCNet.git.
Related papers
- Distractors-Immune Representation Learning with Cross-modal Contrastive Regularization for Change Captioning [71.14084801851381]
Change captioning aims to succinctly describe the semantic change between a pair of similar images.
Most existing methods directly capture the difference between them, which risk obtaining error-prone difference features.
We propose a distractors-immune representation learning network that correlates the corresponding channels of two image representations.
arXiv Detail & Related papers (2024-07-16T13:00:33Z) - Fine-grained Recognition with Learnable Semantic Data Augmentation [68.48892326854494]
Fine-grained image recognition is a longstanding computer vision challenge.
We propose diversifying the training data at the feature-level to alleviate the discriminative region loss problem.
Our method significantly improves the generalization performance on several popular classification networks.
arXiv Detail & Related papers (2023-09-01T11:15:50Z) - iCAR: Bridging Image Classification and Image-text Alignment for Visual
Recognition [33.2800417526215]
Image classification, which classifies images by pre-defined categories, has been the dominant approach to visual representation learning over the last decade.
Visual learning through image-text alignment, however, has emerged to show promising performance, especially for zero-shot recognition.
We propose a deep fusion method with three adaptations that effectively bridge two learning tasks.
arXiv Detail & Related papers (2022-04-22T15:27:21Z) - SGMNet: Scene Graph Matching Network for Few-Shot Remote Sensing Scene
Classification [14.016637774748677]
Few-Shot Remote Sensing Scene Classification (FSRSSC) is an important task, which aims to recognize novel scene classes with few examples.
We propose a novel scene graph matching-based meta-learning framework for FSRSSC, called SGMNet.
We conduct extensive experiments on UCMerced LandUse, WHU19, AID, and NWPU-RESISC45 datasets.
arXiv Detail & Related papers (2021-10-09T07:43:40Z) - Learning Contrastive Representation for Semantic Correspondence [150.29135856909477]
We propose a multi-level contrastive learning approach for semantic matching.
We show that image-level contrastive learning is a key component to encourage the convolutional features to find correspondence between similar objects.
arXiv Detail & Related papers (2021-09-22T18:34:14Z) - Semantic Distribution-aware Contrastive Adaptation for Semantic
Segmentation [50.621269117524925]
Domain adaptive semantic segmentation refers to making predictions on a certain target domain with only annotations of a specific source domain.
We present a semantic distribution-aware contrastive adaptation algorithm that enables pixel-wise representation alignment.
We evaluate SDCA on multiple benchmarks, achieving considerable improvements over existing algorithms.
arXiv Detail & Related papers (2021-05-11T13:21:25Z) - Learning to Focus: Cascaded Feature Matching Network for Few-shot Image
Recognition [38.49419948988415]
Deep networks can learn to accurately recognize objects of a category by training on a large number of images.
A meta-learning challenge known as a low-shot image recognition task comes when only a few images with annotations are available for learning a recognition model for one category.
Our method, called Cascaded Feature Matching Network (CFMN), is proposed to solve this problem.
Experiments for few-shot learning on two standard datasets, emphminiImageNet and Omniglot, have confirmed the effectiveness of our method.
arXiv Detail & Related papers (2021-01-13T11:37:28Z) - Self-Supervised Ranking for Representation Learning [108.38993212650577]
We present a new framework for self-supervised representation learning by formulating it as a ranking problem in an image retrieval context.
We train a representation encoder by maximizing average precision (AP) for ranking, where random views of an image are considered positively related.
In principle, by using a ranking criterion, we eliminate reliance on object-centric curated datasets.
arXiv Detail & Related papers (2020-10-14T17:24:56Z) - Epipolar-Guided Deep Object Matching for Scene Change Detection [23.951526610952765]
This paper describes a viewpoint-robust object-based change detection network (OBJ-CDNet)
Mobile cameras capture images from different viewpoints each time due to differences in camera trajectory and shutter timing.
We introduce a deep graph matching network that establishes object correspondence between an image pair.
arXiv Detail & Related papers (2020-07-30T15:48:40Z) - Learning Attentive Pairwise Interaction for Fine-Grained Classification [53.66543841939087]
We propose a simple but effective Attentive Pairwise Interaction Network (API-Net) for fine-grained classification.
API-Net first learns a mutual feature vector to capture semantic differences in the input pair.
It then compares this mutual vector with individual vectors to generate gates for each input image.
We conduct extensive experiments on five popular benchmarks in fine-grained classification.
arXiv Detail & Related papers (2020-02-24T12:17:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.