Generating Multi-scale Maps from Remote Sensing Images via Series
Generative Adversarial Networks
- URL: http://arxiv.org/abs/2103.16909v1
- Date: Wed, 31 Mar 2021 08:58:37 GMT
- Title: Generating Multi-scale Maps from Remote Sensing Images via Series
Generative Adversarial Networks
- Authors: Xu Chen, Bangguo Yin, Songqiang Chen, Haifeng Li and Tian Xu
- Abstract summary: We develop a series strategy of generators for multi-scale rs2map translation.
High-resolution RSIs are inputted to an rs2map model to output large-scale maps.
Experiments show better quality multi-scale map generation with the series strategy.
- Score: 12.34648824166359
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Considering the success of generative adversarial networks (GANs) for
image-to-image translation, researchers have attempted to translate remote
sensing images (RSIs) to maps (rs2map) through GAN for cartography. However,
these studies involved limited scales, which hinders multi-scale map creation.
By extending their method, multi-scale RSIs can be trivially translated to
multi-scale maps (multi-scale rs2map translation) through scale-wise rs2map
models trained for certain scales (parallel strategy). However, this strategy
has two theoretical limitations. First, inconsistency between various spatial
resolutions of multi-scale RSIs and object generalization on multi-scale maps
(RS-m inconsistency) increasingly complicate the extraction of geographical
information from RSIs for rs2map models with decreasing scale. Second, as
rs2map translation is cross-domain, generators incur high computation costs to
transform the RSI pixel distribution to that on maps. Thus, we designed a
series strategy of generators for multi-scale rs2map translation to address
these limitations. In this strategy, high-resolution RSIs are inputted to an
rs2map model to output large-scale maps, which are translated to multi-scale
maps through series multi-scale map translation models. The series strategy
avoids RS-m inconsistency as inputs are high-resolution large-scale RSIs, and
reduces the distribution gap in multi-scale map generation through similar
pixel distributions among multi-scale maps. Our experimental results showed
better quality multi-scale map generation with the series strategy, as shown by
average increases of 11.69%, 53.78%, 55.42%, and 72.34% in the structural
similarity index, edge structural similarity index, intersection over union
(road), and intersection over union (water) for data from Mexico City and Tokyo
at zoom level 17-13.
Related papers
- MapGlue: Multimodal Remote Sensing Image Matching [12.376931699274062]
Multimodal remote sensing image (MRSI) matching is pivotal for cross-modal fusion, localization, and object detection.
Existing unimodal datasets lack scale and diversity, limiting deep learning solutions.
This paper proposes MapGlue, a universal MRSI matching framework, and MapData, a large-scale multimodal dataset addressing these gaps.
arXiv Detail & Related papers (2025-03-20T14:36:16Z) - C2GM: Cascading conditional generative cartography framework for multi-scale tile map generation with geographic feature constraints [2.414525855161937]
This paper presents C2GM, a novel framework for generating multi-scale tile maps from remote sensing images.
We implement a conditional feature fusion encoder to extract object priors from remote sensing images and cascade reference double branch input.
C2GM consistently achieves the state-of-the-art (SOTA) performance on all metrics.
arXiv Detail & Related papers (2025-02-07T15:11:31Z) - ROI-Aware Multiscale Cross-Attention Vision Transformer for Pest Image
Identification [1.9580473532948401]
We propose a novel ROI-aware multiscale cross-attention vision transformer (ROI-ViT)
The proposed ROI-ViT is designed using dual branches, called Pest and ROI branches, which take different types of maps as input: Pest images and ROI maps.
The experimental results show that the proposed ROI-ViT achieves 81.81%, 99.64%, and 84.66% for IP102, D0, and SauTeg pest datasets, respectively.
arXiv Detail & Related papers (2023-12-28T09:16:27Z) - Learning Continuous Depth Representation via Geometric Spatial
Aggregator [47.1698365486215]
We propose a novel continuous depth representation for depth map super-resolution (DSR)
The heart of this representation is our proposed Geometric Spatial Aggregator (GSA), which exploits a distance field modulated by arbitrarily upsampled target gridding.
We also present a transformer-style backbone named GeoDSR, which possesses a principled way to construct the functional mapping between local coordinates.
arXiv Detail & Related papers (2022-12-07T07:48:23Z) - Hierarchical Transformer for Survival Prediction Using Multimodality
Whole Slide Images and Genomics [63.76637479503006]
Learning good representation of giga-pixel level whole slide pathology images (WSI) for downstream tasks is critical.
This paper proposes a hierarchical-based multimodal transformer framework that learns a hierarchical mapping between pathology images and corresponding genes.
Our architecture requires fewer GPU resources compared with benchmark methods while maintaining better WSI representation ability.
arXiv Detail & Related papers (2022-11-29T23:47:56Z) - Redesigning Multi-Scale Neural Network for Crowd Counting [68.674652984003]
We introduce a hierarchical mixture of density experts, which hierarchically merges multi-scale density maps for crowd counting.
Within the hierarchical structure, an expert competition and collaboration scheme is presented to encourage contributions from all scales.
Experiments show that our method achieves the state-of-the-art performance on five public datasets.
arXiv Detail & Related papers (2022-08-04T21:49:29Z) - Unpaired Image Super-Resolution with Optimal Transport Maps [128.1189695209663]
Real-world image super-resolution (SR) tasks often do not have paired datasets limiting the application of supervised techniques.
We propose an algorithm for unpaired SR which learns an unbiased OT map for the perceptual transport cost.
Our algorithm provides nearly state-of-the-art performance on the large-scale unpaired AIM-19 dataset.
arXiv Detail & Related papers (2022-02-02T16:21:20Z) - Gradient Variance Loss for Structure-Enhanced Image Super-Resolution [16.971608518924597]
We introduce a structure-enhancing loss function, coined Gradient Variance (GV) loss, and generate textures with perceptual-pleasant details.
Experimental results show that the GV loss can significantly improve both Structure Similarity (SSIM) and peak signal-to-noise ratio (PSNR) performance of existing image super-resolution (SR) deep learning models.
arXiv Detail & Related papers (2022-02-02T12:31:05Z) - LocalTrans: A Multiscale Local Transformer Network for Cross-Resolution
Homography Estimation [52.63874513999119]
Cross-resolution image alignment is a key problem in multiscale giga photography.
Existing deep homography methods neglecting the explicit formulation of correspondences between them, which leads to degraded accuracy in cross-resolution challenges.
We propose a local transformer network embedded within a multiscale structure to explicitly learn correspondences between the multimodal inputs.
arXiv Detail & Related papers (2021-06-08T02:51:45Z) - Spatial-spectral Hyperspectral Image Classification via Multiple Random
Anchor Graphs Ensemble Learning [88.60285937702304]
This paper proposes a novel spatial-spectral HSI classification method via multiple random anchor graphs ensemble learning (RAGE)
Firstly, the local binary pattern is adopted to extract the more descriptive features on each selected band, which preserves local structures and subtle changes of a region.
Secondly, the adaptive neighbors assignment is introduced in the construction of anchor graph, to reduce the computational complexity.
arXiv Detail & Related papers (2021-03-25T09:31:41Z) - ASDN: A Deep Convolutional Network for Arbitrary Scale Image
Super-Resolution [6.672537252929814]
Image viewer applications commonly allow users to zoom the images to arbitrary magnification scales.
This paper employs a Laplacian pyramid method to reconstruct any-scale high-resolution (HR) images using the high-frequency image details.
arXiv Detail & Related papers (2020-10-06T01:18:46Z) - Sequential Hierarchical Learning with Distribution Transformation for
Image Super-Resolution [83.70890515772456]
We build a sequential hierarchical learning super-resolution network (SHSR) for effective image SR.
We consider the inter-scale correlations of features, and devise a sequential multi-scale block (SMB) to progressively explore the hierarchical information.
Experiment results show SHSR achieves superior quantitative performance and visual quality to state-of-the-art methods.
arXiv Detail & Related papers (2020-07-19T01:35:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.