DouFu: A Double Fusion Joint Learning Method For Driving Trajectory
Representation
- URL: http://arxiv.org/abs/2205.08356v2
- Date: Fri, 14 Oct 2022 05:11:51 GMT
- Title: DouFu: A Double Fusion Joint Learning Method For Driving Trajectory
Representation
- Authors: Han Wang, Zhou Huang, Xiao Zhou, Ganmin Yin, Yi Bao, Yi Zhang
- Abstract summary: We propose a novel multimodal fusion model, DouFu, for trajectory representation joint learning.
We first design movement, route, and global features generated from the trajectory data and urban functional zones.
With the global semantic feature, DouFu produces a comprehensive embedding for each trajectory.
- Score: 13.321587117066166
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Driving trajectory representation learning is of great significance for
various location-based services, such as driving pattern mining and route
recommendation. However, previous representation generation approaches tend to
rarely address three challenges: 1) how to represent the intricate semantic
intentions of mobility inexpensively; 2) complex and weak spatial-temporal
dependencies due to the sparsity and heterogeneity of the trajectory data; 3)
route selection preferences and their correlation to driving behavior. In this
paper, we propose a novel multimodal fusion model, DouFu, for trajectory
representation joint learning, which applies multimodal learning and attention
fusion module to capture the internal characteristics of trajectories. We first
design movement, route, and global features generated from the trajectory data
and urban functional zones and then analyze them respectively with the
attention encoder or feed forward network. The attention fusion module
incorporates route features with movement features to create a better
spatial-temporal embedding. With the global semantic feature, DouFu produces a
comprehensive embedding for each trajectory. We evaluate representations
generated by our method and other baseline models on classification and
clustering tasks. Empirical results show that DouFu outperforms other models in
most of the learning algorithms like the linear regression and the support
vector machine by more than 10%.
Related papers
- Context-Enhanced Multi-View Trajectory Representation Learning: Bridging the Gap through Self-Supervised Models [27.316692263196277]
MVTraj is a novel multi-view modeling method for trajectory representation learning.
It integrates diverse contextual knowledge, from GPS to road network and points-of-interest to provide a more comprehensive understanding of trajectory data.
Extensive experiments on real-world datasets demonstrate that MVTraj significantly outperforms existing baselines in tasks associated with various spatial views.
arXiv Detail & Related papers (2024-10-17T03:56:12Z) - Micro-Macro Spatial-Temporal Graph-based Encoder-Decoder for Map-Constrained Trajectory Recovery [21.911875343270683]
Recovering missing GPS points in a sparse trajectory could offer deep insights into users' moving behaviors in intelligent transportation systems.
It is extremely hard for them to comprehensively capture the micro-semantics of individual trajectory.
We propose a Micro-Macro- Spatial Graph-Decoder (MM-STGED) to efficiently describe the micro-semantics of trajectory and design a novel message-passing mechanism.
arXiv Detail & Related papers (2024-04-29T22:54:35Z) - More Than Routing: Joint GPS and Route Modeling for Refine Trajectory
Representation Learning [26.630640299709114]
We propose Joint GPS and Route Modelling based on self-supervised technology, namely JGRM.
We develop two encoders, each tailored to capture representations of route and GPS trajectories respectively.
The representations from the two modalities are fed into a shared transformer for inter-modal information interaction.
arXiv Detail & Related papers (2024-02-25T18:27:25Z) - Rethinking Range View Representation for LiDAR Segmentation [66.73116059734788]
"Many-to-one" mapping, semantic incoherence, and shape deformation are possible impediments against effective learning from range view projections.
We present RangeFormer, a full-cycle framework comprising novel designs across network architecture, data augmentation, and post-processing.
We show that, for the first time, a range view method is able to surpass the point, voxel, and multi-view fusion counterparts in the competing LiDAR semantic and panoptic segmentation benchmarks.
arXiv Detail & Related papers (2023-03-09T16:13:27Z) - Hybrid Routing Transformer for Zero-Shot Learning [83.64532548391]
This paper presents a novel transformer encoder-decoder model, called hybrid routing transformer (HRT)
We embed an active attention, which is constructed by both the bottom-up and the top-down dynamic routing pathways to generate the attribute-aligned visual feature.
While in HRT decoder, we use static routing to calculate the correlation among the attribute-aligned visual features, the corresponding attribute semantics, and the class attribute vectors to generate the final class label predictions.
arXiv Detail & Related papers (2022-03-29T07:55:08Z) - Aerial Images Meet Crowdsourced Trajectories: A New Approach to Robust
Road Extraction [110.61383502442598]
We introduce a novel neural network framework termed Cross-Modal Message Propagation Network (CMMPNet)
CMMPNet is composed of two deep Auto-Encoders for modality-specific representation learning and a tailor-designed Dual Enhancement Module for cross-modal representation refinement.
Experiments on three real-world benchmarks demonstrate the effectiveness of our CMMPNet for robust road extraction.
arXiv Detail & Related papers (2021-11-30T04:30:10Z) - Multi-Modal Fusion Transformer for End-to-End Autonomous Driving [59.60483620730437]
We propose TransFuser, a novel Multi-Modal Fusion Transformer, to integrate image and LiDAR representations using attention.
Our approach achieves state-of-the-art driving performance while reducing collisions by 76% compared to geometry-based fusion.
arXiv Detail & Related papers (2021-04-19T11:48:13Z) - A Driving Behavior Recognition Model with Bi-LSTM and Multi-Scale CNN [59.57221522897815]
We propose a neural network model based on trajectories information for driving behavior recognition.
We evaluate the proposed model on the public BLVD dataset, achieving a satisfying performance.
arXiv Detail & Related papers (2021-03-01T06:47:29Z) - Multi-intersection Traffic Optimisation: A Benchmark Dataset and a
Strong Baseline [85.9210953301628]
Control of traffic signals is fundamental and critical to alleviate traffic congestion in urban areas.
Because of the high complexity of modelling the problem, experimental settings of current works are often inconsistent.
We propose a novel and strong baseline model based on deep reinforcement learning with the encoder-decoder structure.
arXiv Detail & Related papers (2021-01-24T03:55:39Z) - Mobility Inference on Long-Tailed Sparse Trajectory [2.4444287331956898]
We propose a single trajectory inference algorithm that utilizes a generic long-tailed sparsity pattern in the large-scale trajectory data.
The algorithm guarantees a 100% precision in the stay/travel inference with a provable lower-bound in the recall.
Evaluations with three trajectory data sets of 40 million urban users validate the performance guarantees of the proposed inference algorithm.
arXiv Detail & Related papers (2020-01-21T16:32:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.