Inductive Graph Transformer for Delivery Time Estimation
- URL: http://arxiv.org/abs/2211.02863v1
- Date: Sat, 5 Nov 2022 09:51:15 GMT
- Title: Inductive Graph Transformer for Delivery Time Estimation
- Authors: Xin Zhou, Jinglong Wang, Yong Liu, Xingyu Wu, Zhiqi Shen, Cyril Leung
- Abstract summary: We propose an inductive graph transformer (IGT) that leverages raw feature information and structural graph data to estimate package delivery time.
Experiments on real-world logistics datasets show that our proposed model can significantly outperform the state-of-the-art methods on estimation of delivery time.
- Score: 19.024006381947416
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Providing accurate estimated time of package delivery on users' purchasing
pages for e-commerce platforms is of great importance to their purchasing
decisions and post-purchase experiences. Although this problem shares some
common issues with the conventional estimated time of arrival (ETA), it is more
challenging with the following aspects: 1) Inductive inference. Models are
required to predict ETA for orders with unseen retailers and addresses; 2)
High-order interaction of order semantic information. Apart from the
spatio-temporal features, the estimated time also varies greatly with other
factors, such as the packaging efficiency of retailers, as well as the
high-order interaction of these factors. In this paper, we propose an inductive
graph transformer (IGT) that leverages raw feature information and structural
graph data to estimate package delivery time. Different from previous graph
transformer architectures, IGT adopts a decoupled pipeline and trains
transformer as a regression function that can capture the multiplex information
from both raw feature and dense embeddings encoded by a graph neural network
(GNN). In addition, we further simplify the GNN structure by removing its
non-linear activation and the learnable linear transformation matrix. The
reduced parameter search space and linear information propagation in the
simplified GNN enable the IGT to be applied in large-scale industrial
scenarios. Experiments on real-world logistics datasets show that our proposed
model can significantly outperform the state-of-the-art methods on estimation
of delivery time. The source code is available at:
https://github.com/enoche/IGT-WSDM23.
Related papers
- Deep Prompt Tuning for Graph Transformers [55.2480439325792]
Fine-tuning is resource-intensive and requires storing multiple copies of large models.
We propose a novel approach called deep graph prompt tuning as an alternative to fine-tuning.
By freezing the pre-trained parameters and only updating the added tokens, our approach reduces the number of free parameters and eliminates the need for multiple model copies.
arXiv Detail & Related papers (2023-09-18T20:12:17Z) - TransformerG2G: Adaptive time-stepping for learning temporal graph
embeddings using transformers [2.2120851074630177]
We develop a graph embedding model with uncertainty quantification, TransformerG2G, to learn temporal dynamics of temporal graphs.
Our experiments demonstrate that the proposed TransformerG2G model outperforms conventional multi-step methods.
By examining the attention weights, we can uncover temporal dependencies, identify influential elements, and gain insights into the complex interactions within the graph structure.
arXiv Detail & Related papers (2023-07-05T18:34:22Z) - Networked Time Series Imputation via Position-aware Graph Enhanced
Variational Autoencoders [31.953958053709805]
We design a new model named PoGeVon which leverages variational autoencoder (VAE) to predict missing values over both node time series features and graph structures.
Experiment results demonstrate the effectiveness of our model over baselines.
arXiv Detail & Related papers (2023-05-29T21:11:34Z) - G-Adapter: Towards Structure-Aware Parameter-Efficient Transfer Learning
for Graph Transformer Networks [0.7118812771905295]
We show that it is sub-optimal to directly transfer existing PEFTs to graph-based tasks due to the issue of feature distribution shift.
We propose a novel structure-aware PEFT approach, named G-Adapter, to guide the updating process.
Extensive experiments demonstrate that G-Adapter obtains the state-of-the-art performance compared to the counterparts on nine graph benchmark datasets.
arXiv Detail & Related papers (2023-05-17T16:10:36Z) - EasyDGL: Encode, Train and Interpret for Continuous-time Dynamic Graph Learning [92.71579608528907]
This paper aims to design an easy-to-use pipeline (termed as EasyDGL) composed of three key modules with both strong ability fitting and interpretability.
EasyDGL can effectively quantify the predictive power of frequency content that a model learn from the evolving graph data.
arXiv Detail & Related papers (2023-03-22T06:35:08Z) - Graph Decision Transformer [83.76329715043205]
Graph Decision Transformer (GDT) is a novel offline reinforcement learning approach.
GDT models the input sequence into a causal graph to capture potential dependencies between fundamentally different concepts.
Our experiments show that GDT matches or surpasses the performance of state-of-the-art offline RL methods on image-based Atari and OpenAI Gym.
arXiv Detail & Related papers (2023-03-07T09:10:34Z) - FormerTime: Hierarchical Multi-Scale Representations for Multivariate
Time Series Classification [53.55504611255664]
FormerTime is a hierarchical representation model for improving the classification capacity for the multivariate time series classification task.
It exhibits three aspects of merits: (1) learning hierarchical multi-scale representations from time series data, (2) inheriting the strength of both transformers and convolutional networks, and (3) tacking the efficiency challenges incurred by the self-attention mechanism.
arXiv Detail & Related papers (2023-02-20T07:46:14Z) - Towards Flexible Inductive Bias via Progressive Reparameterization
Scheduling [25.76814731638375]
There are two de facto standard architectures in computer vision: Convolutional Neural Networks (CNNs) and Vision Transformers (ViTs)
We show these approaches overlook that the optimal inductive bias also changes according to the target data scale changes.
The more convolution-like inductive bias is included in the model, the smaller the data scale is required where the ViT-like model outperforms the ResNet performance.
arXiv Detail & Related papers (2022-10-04T04:20:20Z) - Unboxing the graph: Neural Relational Inference for Mobility Prediction [15.4049962498675]
Graph Networks (GNNs) have been widely applied on non-euclidean spatial data.
In this paper, we use Neural Inference, a dynamically learn the optimal graph model.
arXiv Detail & Related papers (2022-01-25T13:26:35Z) - Connecting the Dots: Multivariate Time Series Forecasting with Graph
Neural Networks [91.65637773358347]
We propose a general graph neural network framework designed specifically for multivariate time series data.
Our approach automatically extracts the uni-directed relations among variables through a graph learning module.
Our proposed model outperforms the state-of-the-art baseline methods on 3 of 4 benchmark datasets.
arXiv Detail & Related papers (2020-05-24T04:02:18Z) - Transformer Hawkes Process [79.16290557505211]
We propose a Transformer Hawkes Process (THP) model, which leverages the self-attention mechanism to capture long-term dependencies.
THP outperforms existing models in terms of both likelihood and event prediction accuracy by a notable margin.
We provide a concrete example, where THP achieves improved prediction performance for learning multiple point processes when incorporating their relational information.
arXiv Detail & Related papers (2020-02-21T13:48:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.