Boundary Prompting: Elastic Urban Region Representation via Graph-based Spatial Tokenization
- URL: http://arxiv.org/abs/2503.07991v1
- Date: Tue, 11 Mar 2025 02:49:58 GMT
- Title: Boundary Prompting: Elastic Urban Region Representation via Graph-based Spatial Tokenization
- Authors: Haojia Zhu, Jiahui Jin, Dong Kan, Rouxi Shen, Ruize Wang, Xiangguo Sun, Jinghui Zhang,
- Abstract summary: Boundary Prompting Urban Region Representation Framework (BPURF) is a novel approach that allows for elastic urban region definitions.<n>BPURF comprises two key components: a spatial token dictionary and a region token set representation model.<n>This framework enables the definition of urban regions through boundary prompting, supporting varying region boundaries and adapting to different tasks.
- Score: 4.638735746777273
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Urban region representation is essential for various applications such as urban planning, resource allocation, and policy development. Traditional methods rely on fixed, predefined region boundaries, which fail to capture the dynamic and complex nature of real-world urban areas. In this paper, we propose the Boundary Prompting Urban Region Representation Framework (BPURF), a novel approach that allows for elastic urban region definitions. BPURF comprises two key components: (1) A spatial token dictionary, where urban entities are treated as tokens and integrated into a unified token graph, and (2) a region token set representation model which utilize token aggregation and a multi-channel model to embed token sets corresponding to region boundaries. Additionally, we propose fast token set extraction strategy to enable online token set extraction during training and prompting. This framework enables the definition of urban regions through boundary prompting, supporting varying region boundaries and adapting to different tasks. Extensive experiments demonstrate the effectiveness of BPURF in capturing the complex characteristics of urban regions.
Related papers
- Urban Region Representation Learning: A Flexible Approach [14.345157663791133]
We propose a model named FlexiReg for urban region representation learning that is flexible with both the formation of urban regions and the input region features.
We show that FlexiReg outperforms state-of-the-art models by up to 202% in term of the accuracy of four diverse downstream tasks using the produced urban region representations.
arXiv Detail & Related papers (2025-03-12T07:33:44Z) - Origin-Destination Demand Prediction: An Urban Radiation and Attraction Perspective [47.472045876664424]
We present a new model that captures the relationships between two types of capacities.
Specifically, we first model regions' radiation and attraction capacities using a bilateral branch network.
We then describe the transformation relationship of different capacities of the same region using a hypergraph-based parameter generation method.
arXiv Detail & Related papers (2024-11-29T15:35:17Z) - Region-Aware Text-to-Image Generation via Hard Binding and Soft Refinement [40.94329069897935]
We present RAG, a Regional-Aware text-to-image Generation method conditioned on regional descriptions for precise layout composition.
RAG achieves superior performance over attribute binding and object relationship than previous tuning-free methods.
arXiv Detail & Related papers (2024-11-10T18:45:41Z) - Urban Region Pre-training and Prompting: A Graph-based Approach [10.375941950028938]
We propose a $textbfG$raph-based $textbfU$rban $textbfR$egion $textbfP$re-training and $textbfP$rompting framework for region representation learning.
arXiv Detail & Related papers (2024-08-12T05:00:23Z) - Differentiable Reasoning about Knowledge Graphs with Region-based Graph Neural Networks [62.93577376960498]
Methods for knowledge graph (KG) completion need to capture semantic regularities and use these regularities to infer plausible knowledge that is not explicitly stated.
Most embedding-based methods are opaque in the kinds of regularities they can capture, although region-based KG embedding models have emerged as a more transparent alternative.
We propose RESHUFFLE, a simple model based on ordering constraints that can faithfully capture a much larger class of rule bases than existing approaches.
arXiv Detail & Related papers (2024-06-13T18:37:24Z) - RegionGPT: Towards Region Understanding Vision Language Model [88.42271128373191]
RegionGPT (short as RGPT) is a novel framework designed for complex region-level captioning and understanding.
We develop an automated region caption data generation pipeline, enriching the training set with detailed region-level captions.
We demonstrate that a universal RGPT model can be effectively applied and significantly enhancing performance across a range of region-level tasks.
arXiv Detail & Related papers (2024-03-04T18:58:08Z) - Cross-City Matters: A Multimodal Remote Sensing Benchmark Dataset for
Cross-City Semantic Segmentation using High-Resolution Domain Adaptation
Networks [82.82866901799565]
We build a new set of multimodal remote sensing benchmark datasets (including hyperspectral, multispectral, SAR) for the study purpose of the cross-city semantic segmentation task.
Beyond the single city, we propose a high-resolution domain adaptation network, HighDAN, to promote the AI model's generalization ability from the multi-city environments.
HighDAN is capable of retaining the spatially topological structure of the studied urban scene well in a parallel high-to-low resolution fusion fashion.
arXiv Detail & Related papers (2023-09-26T23:55:39Z) - Attentive Graph Enhanced Region Representation Learning [7.4106801792345705]
Representing urban regions accurately and comprehensively is essential for various urban planning and analysis tasks.
We propose the Attentive Graph Enhanced Region Representation Learning (ATGRL) model, which aims to capture comprehensive dependencies from multiple graphs and learn rich semantic representations of urban regions.
arXiv Detail & Related papers (2023-07-06T16:38:43Z) - Semantic Segmentation by Early Region Proxy [53.594035639400616]
We present a novel and efficient modeling that starts from interpreting the image as a tessellation of learnable regions.
To model region-wise context, we exploit Transformer to encode regions in a sequence-to-sequence manner.
Semantic segmentation is now carried out as per-region prediction on top of the encoded region embeddings.
arXiv Detail & Related papers (2022-03-26T10:48:32Z) - Urban Region Profiling via A Multi-Graph Representation Learning
Framework [0.0]
We propose a multi-graph representative learning framework, called Region2Vec, for urban region profiling.
Experiments on real-world datasets show that Region2Vec can be employed in three applications and outperforms all state-of-the-art baselines.
arXiv Detail & Related papers (2022-02-04T11:05:37Z) - Learning Neighborhood Representation from Multi-Modal Multi-Graph:
Image, Text, Mobility Graph and Beyond [20.014906526266795]
We propose a novel approach to integrate multi-modal geotagged inputs as either node or edge features of a multi-graph.
Specifically, we use street view images and POI features to characterize neighborhoods (nodes) and use human mobility to characterize the relationship between neighborhoods (directed edges)
The embedding we trained outperforms the ones using only unimodal data as regional inputs.
arXiv Detail & Related papers (2021-05-06T07:44:05Z) - Methodological Foundation of a Numerical Taxonomy of Urban Form [62.997667081978825]
We present a method for numerical taxonomy of urban form derived from biological systematics.
We derive homogeneous urban tissue types and, by determining overall morphological similarity between them, generate a hierarchical classification of urban form.
After framing and presenting the method, we test it on two cities - Prague and Amsterdam.
arXiv Detail & Related papers (2021-04-30T12:47:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.