SGMNet: Scene Graph Matching Network for Few-Shot Remote Sensing Scene
Classification
- URL: http://arxiv.org/abs/2110.04494v1
- Date: Sat, 9 Oct 2021 07:43:40 GMT
- Title: SGMNet: Scene Graph Matching Network for Few-Shot Remote Sensing Scene
Classification
- Authors: Baoquan Zhang, Shanshan Feng, Xutao Li, Yunming Ye, and Rui Ye
- Abstract summary: Few-Shot Remote Sensing Scene Classification (FSRSSC) is an important task, which aims to recognize novel scene classes with few examples.
We propose a novel scene graph matching-based meta-learning framework for FSRSSC, called SGMNet.
We conduct extensive experiments on UCMerced LandUse, WHU19, AID, and NWPU-RESISC45 datasets.
- Score: 14.016637774748677
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Few-Shot Remote Sensing Scene Classification (FSRSSC) is an important task,
which aims to recognize novel scene classes with few examples. Recently,
several studies attempt to address the FSRSSC problem by following few-shot
natural image classification methods. These existing methods have made
promising progress and achieved superior performance. However, they all
overlook two unique characteristics of remote sensing images: (i) object
co-occurrence that multiple objects tend to appear together in a scene image
and (ii) object spatial correlation that these co-occurrence objects are
distributed in the scene image following some spatial structure patterns. Such
unique characteristics are very beneficial for FSRSSC, which can effectively
alleviate the scarcity issue of labeled remote sensing images since they can
provide more refined descriptions for each scene class. To fully exploit these
characteristics, we propose a novel scene graph matching-based meta-learning
framework for FSRSSC, called SGMNet. In this framework, a scene graph
construction module is carefully designed to represent each test remote sensing
image or each scene class as a scene graph, where the nodes reflect these
co-occurrence objects meanwhile the edges capture the spatial correlations
between these co-occurrence objects. Then, a scene graph matching module is
further developed to evaluate the similarity score between each test remote
sensing image and each scene class. Finally, based on the similarity scores, we
perform the scene class prediction via a nearest neighbor classifier. We
conduct extensive experiments on UCMerced LandUse, WHU19, AID, and
NWPU-RESISC45 datasets. The experimental results show that our method obtains
superior performance over the previous state-of-the-art methods.
Related papers
- Multi-level Cross-modal Feature Alignment via Contrastive Learning
towards Zero-shot Classification of Remote Sensing Image Scenes [7.17717863134783]
Cross-modal feature alignment methods have been proposed to address the zero-shot image scene classification.
We propose a multi-level cross-modal feature alignment method via contrastive learning for zero-shot classification of remote sensing image scenes.
Our proposed method outperforms state of the art methods for zero-shot remote sensing image scene classification.
arXiv Detail & Related papers (2023-05-31T10:00:45Z) - Self-attention on Multi-Shifted Windows for Scene Segmentation [14.47974086177051]
We explore the effective use of self-attention within multi-scale image windows to learn descriptive visual features.
We propose three different strategies to aggregate these feature maps to decode the feature representation for dense prediction.
Our models achieve very promising performance on four public scene segmentation datasets.
arXiv Detail & Related papers (2022-07-10T07:36:36Z) - Complex Scene Image Editing by Scene Graph Comprehension [17.72638225034884]
We propose a two-stage method for achieving complex scene image editing by Scene Graph (SGC-Net)
In the first stage, we train a Region of Interest (RoI) prediction network that uses scene graphs and predict the locations of the target objects.
The second stage uses a conditional diffusion model to edit the image based on our RoI predictions.
arXiv Detail & Related papers (2022-03-24T05:12:54Z) - Robust Region Feature Synthesizer for Zero-Shot Object Detection [87.79902339984142]
We build a novel zero-shot object detection framework that contains an Intra-class Semantic Diverging component and an Inter-class Structure Preserving component.
It is the first study to carry out zero-shot object detection in remote sensing imagery.
arXiv Detail & Related papers (2022-01-01T03:09:15Z) - Rectifying the Shortcut Learning of Background: Shared Object
Concentration for Few-Shot Image Recognition [101.59989523028264]
Few-Shot image classification aims to utilize pretrained knowledge learned from a large-scale dataset to tackle a series of downstream classification tasks.
We propose COSOC, a novel Few-Shot Learning framework, to automatically figure out foreground objects at both pretraining and evaluation stage.
arXiv Detail & Related papers (2021-07-16T07:46:41Z) - DetCo: Unsupervised Contrastive Learning for Object Detection [64.22416613061888]
Unsupervised contrastive learning achieves great success in learning image representations with CNN.
We present a novel contrastive learning approach, named DetCo, which fully explores the contrasts between global image and local image patches.
DetCo consistently outperforms supervised method by 1.6/1.2/1.0 AP on Mask RCNN-C4/FPN/RetinaNet with 1x schedule.
arXiv Detail & Related papers (2021-02-09T12:47:20Z) - Non-parametric spatially constrained local prior for scene parsing on
real-world data [4.035753155957698]
We present the non-parametric Spatially Constrained Local Prior (SCLP) for scene parsing on realistic data.
The SCLP is powerful in capturing both long- and short-range context about inter-object correlations in the query image.
Our experiments on the SIFT Flow and PASCAL-Context benchmark datasets show that the SCLP used in conjunction with superpixel-level visual features achieves one of the top performance compared with state-of-the-art approaches.
arXiv Detail & Related papers (2020-06-23T10:12:08Z) - Gradient-Induced Co-Saliency Detection [81.54194063218216]
Co-saliency detection (Co-SOD) aims to segment the common salient foreground in a group of relevant images.
In this paper, inspired by human behavior, we propose a gradient-induced co-saliency detection method.
arXiv Detail & Related papers (2020-04-28T08:40:55Z) - High-Order Information Matters: Learning Relation and Topology for
Occluded Person Re-Identification [84.43394420267794]
We propose a novel framework by learning high-order relation and topology information for discriminative features and robust alignment.
Our framework significantly outperforms state-of-the-art by6.5%mAP scores on Occluded-Duke dataset.
arXiv Detail & Related papers (2020-03-18T12:18:35Z) - Improving Few-shot Learning by Spatially-aware Matching and
CrossTransformer [116.46533207849619]
We study the impact of scale and location mismatch in the few-shot learning scenario.
We propose a novel Spatially-aware Matching scheme to effectively perform matching across multiple scales and locations.
arXiv Detail & Related papers (2020-01-06T14:10:20Z) - Scene Graph Generation via Conditional Random Fields [14.282277071380447]
We propose a novel scene graph generation model for predicting object instances and its corresponding relationships in an image.
Our model, SG-CRF, learns the sequential order of subject and object in a relationship triplet, and the semantic compatibility of object nodes instance and relationship nodes in a scene graph efficiently.
arXiv Detail & Related papers (2018-11-20T04:55:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.