Joint predictions of multi-modal ride-hailing demands: a deep multi-task
multigraph learning-based approach
- URL: http://arxiv.org/abs/2011.05602v1
- Date: Wed, 11 Nov 2020 07:10:50 GMT
- Title: Joint predictions of multi-modal ride-hailing demands: a deep multi-task
multigraph learning-based approach
- Authors: Jintao Ke, Siyuan Feng, Zheng Zhu, Hai Yang, Jieping Ye
- Abstract summary: We propose a deep multi-task multi-graph learning approach, which combines multiple multi-graph convolutional (MGC) networks for predicting demands for different service modes.
We show that our propose approach outperforms the benchmark algorithms in prediction accuracy for different ride-hailing modes.
- Score: 64.18639899347822
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Ride-hailing platforms generally provide various service options to
customers, such as solo ride services, shared ride services, etc. It is
generally expected that demands for different service modes are correlated, and
the prediction of demand for one service mode can benefit from historical
observations of demands for other service modes. Moreover, an accurate joint
prediction of demands for multiple service modes can help the platforms better
allocate and dispatch vehicle resources. Although there is a large stream of
literature on ride-hailing demand predictions for one specific service mode,
little efforts have been paid towards joint predictions of ride-hailing demands
for multiple service modes. To address this issue, we propose a deep multi-task
multi-graph learning approach, which combines two components: (1) multiple
multi-graph convolutional (MGC) networks for predicting demands for different
service modes, and (2) multi-task learning modules that enable knowledge
sharing across multiple MGC networks. More specifically, two multi-task
learning structures are established. The first one is the regularized
cross-task learning, which builds cross-task connections among the inputs and
outputs of multiple MGC networks. The second one is the multi-linear
relationship learning, which imposes a prior tensor normal distribution on the
weights of various MGC networks. Although there are no concrete bridges between
different MGC networks, the weights of these networks are constrained by each
other and subject to a common prior distribution. Evaluated with the
for-hire-vehicle datasets in Manhattan, we show that our propose approach
outperforms the benchmark algorithms in prediction accuracy for different
ride-hailing modes.
Related papers
- Multi-modal Semantic Understanding with Contrastive Cross-modal Feature
Alignment [11.897888221717245]
This paper proposes a novel CLIP-guided contrastive-learning-based architecture to perform multi-modal feature alignment.
Our model is simple to implement without using task-specific external knowledge, and thus can easily migrate to other multi-modal tasks.
arXiv Detail & Related papers (2024-03-11T01:07:36Z) - OmniVec: Learning robust representations with cross modal sharing [28.023214572340336]
We present an approach to learn multiple tasks, in multiple modalities, with a unified architecture.
The proposed network is composed of task specific encoders, a common trunk in the middle, followed by task specific prediction heads.
We train the network on all major modalities, e.g. visual, audio, text and 3D, and report results on $22$ diverse and challenging public benchmarks.
arXiv Detail & Related papers (2023-11-07T14:00:09Z) - HiNet: Novel Multi-Scenario & Multi-Task Learning with Hierarchical Information Extraction [50.40732146978222]
Multi-scenario & multi-task learning has been widely applied to many recommendation systems in industrial applications.
We propose a Hierarchical information extraction Network (HiNet) for multi-scenario and multi-task recommendation.
HiNet achieves a new state-of-the-art performance and significantly outperforms existing solutions.
arXiv Detail & Related papers (2023-03-10T17:24:41Z) - OFASys: A Multi-Modal Multi-Task Learning System for Building Generalist
Models [72.8156832931841]
Generalist models are capable of performing diverse multi-modal tasks in a task-agnostic way within a single model.
We release a generalist model learning system, OFASys, built on top of a declarative task interface named multi-modal instruction.
arXiv Detail & Related papers (2022-12-08T17:07:09Z) - Cross-Mode Knowledge Adaptation for Bike Sharing Demand Prediction using
Domain-Adversarial Graph Neural Networks [8.695763084463055]
This study proposes a domain-adversarial multi-relational graph neural network (DA-MRGNN) for bike sharing demand prediction.
A temporal adversarial adaptation network is introduced to extract shareable features from patterns demand of different modes.
Experiments are conducted using real-world bike sharing, subway and ride-hailing data from New York City.
arXiv Detail & Related papers (2022-11-16T13:35:32Z) - Channel Exchanging Networks for Multimodal and Multitask Dense Image
Prediction [125.18248926508045]
We propose Channel-Exchanging-Network (CEN) which is self-adaptive, parameter-free, and more importantly, applicable for both multimodal fusion and multitask learning.
CEN dynamically exchanges channels betweenworks of different modalities.
For the application of dense image prediction, the validity of CEN is tested by four different scenarios.
arXiv Detail & Related papers (2021-12-04T05:47:54Z) - Routing with Self-Attention for Multimodal Capsule Networks [108.85007719132618]
We present a new multimodal capsule network that allows us to leverage the strength of capsules in the context of a multimodal learning framework.
To adapt the capsules to large-scale input data, we propose a novel routing by self-attention mechanism that selects relevant capsules.
This allows not only for robust training with noisy video data, but also to scale up the size of the capsule network compared to traditional routing methods.
arXiv Detail & Related papers (2021-12-01T19:01:26Z) - VLM: Task-agnostic Video-Language Model Pre-training for Video
Understanding [78.28397557433544]
We present a task-agnostic multi-modal pre-training approach that can accept either video or text input, or both for a variety of end tasks.
Experimental results show strong performance across a wider range of tasks than any previous methods, often outperforming task-specific pre-training.
arXiv Detail & Related papers (2021-05-20T19:13:27Z) - Adversarial Multimodal Representation Learning for Click-Through Rate
Prediction [16.10640369157054]
We propose a novel Multimodal Adversarial Representation Network (MARN) for the Click-Through Rate (CTR) prediction task.
A multimodal attention network first calculates the weights of multiple modalities for each item according to its modality-specific features.
A multimodal adversarial network learns modality-in representations where a double-discriminators strategy is introduced.
arXiv Detail & Related papers (2020-03-07T15:50:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.