Beyond the Next Port: A Multi-Task Transformer for Forecasting Future Voyage Segment Durations
- URL: http://arxiv.org/abs/2601.08013v1
- Date: Mon, 12 Jan 2026 21:32:05 GMT
- Title: Beyond the Next Port: A Multi-Task Transformer for Forecasting Future Voyage Segment Durations
- Authors: Nairui Liu, Fang He, Xindi Tang,
- Abstract summary: The study reformulates future-port ETA prediction as a segment-level time-series forecasting problem.<n>We develop a transformer-based architecture that integrates historical sailing durations, destination port congestion, and static vessel descriptors.<n>The proposed framework employs a causally masked attention mechanism to capture long-range temporal dependencies and a multi-task learning head to jointly predict segment sailing durations and port congestion states.
- Score: 3.9443085703523706
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Accurate forecasts of segment-level sailing durations are fundamental to enhancing maritime schedule reliability and optimizing long-term port operations. However, conventional estimated time of arrival (ETA) models are primarily designed for the immediate next port of call and rely heavily on real-time automatic identification system (AIS) data, which is inherently unavailable for future voyage segments. To address this gap, the study reformulates future-port ETA prediction as a segment-level time-series forecasting problem. We develop a transformer-based architecture that integrates historical sailing durations, destination port congestion proxies, and static vessel descriptors. The proposed framework employs a causally masked attention mechanism to capture long-range temporal dependencies and a multi-task learning head to jointly predict segment sailing durations and port congestion states, leveraging shared latent signals to mitigate high uncertainty. Evaluation on a real-world global dataset from 2021 demonstrates the proposed model consistently outperforms a comprehensive suite of competitive baselines. The result shows a relative reduction of 4.85% in mean absolute error (MAE) and 4.95% in mean absolute percentage error (MAPE) compared with sequence baseline models. The relative reductions with gradient boosting machines are 9.39% in MAE and 52.97% in MAPE. Case studies for the major destination port further illustrate the model's superior accuracy.
Related papers
- Uncertainty-Aware Delivery Delay Duration Prediction via Multi-Task Deep Learning [11.2212153491325]
This paper introduces a multi-task deep learning model for delivery delay duration prediction in the presence of significant imbalanced data.<n>The proposed model is evaluated on a large-scale real-world dataset from an industrial partner.<n> Experimental results show that the proposed method achieves a mean absolute error of 0.67-0.91 days for delayed-shipment predictions.
arXiv Detail & Related papers (2026-02-23T19:01:03Z) - Agentic World Modeling for 6G: Near-Real-Time Generative State-Space Reasoning [70.56067503630486]
We argue that sixth-generation (6G) intelligence is not fluent token prediction but calibrated the capacity to imagine and choose.<n>We show that WM-MS3M cuts mean absolute error (MAE) by 1.69% versus MS3M with 32% fewer parameters and similar latency, and achieves 35-80% lower root mean squared error (RMSE) than attention/hybrid baselines with 2.3-4.1x faster inference.
arXiv Detail & Related papers (2025-11-04T17:22:22Z) - LTMSformer: A Local Trend-Aware Attention and Motion State Encoding Transformer for Multi-Agent Trajectory Prediction [6.520837230073969]
We propose a lightweight framework, LTMSformer, to extract temporal-spatial interaction features for trajectory prediction.<n>We also build a Motion State to incorporate high-order motion state attributes, such as acceleration, jerk, heading, etc.<n>Experiment results show that our method outperforms the baseline HiVT-64, reducing the minADE by approximately 4.35%, the minFDE by 8.74%, and the MR by 20%.
arXiv Detail & Related papers (2025-07-07T03:33:14Z) - Temporal-IRL: Modeling Port Congestion and Berth Scheduling with Inverse Reinforcement Learning [13.45559229405042]
The model learns berth scheduling to predict vessel sequencing at the terminal and estimate vessel port stay.<n>We trained and tested the model, achieving demonstrably excellent results.
arXiv Detail & Related papers (2025-06-24T17:59:12Z) - EMForecaster: A Deep Learning Framework for Time Series Forecasting in Wireless Networks with Distribution-Free Uncertainty Quantification [11.88654374102813]
We develop a deep learning (DL) time series forecasting framework referred to as textitEMForecaster.<n> EMForecaster achieves superior performance across diverse datasets, spanning both short-term and long-term prediction horizons.<n>In point forecasting tasks, EMForecaster substantially outperforms current state-of-the-art DL approaches, showing improvements of 53.97% over the Transformer architecture and 38.44% over the average of all baseline models.
arXiv Detail & Related papers (2025-03-31T18:10:08Z) - STAA: Spatio-Temporal Alignment Attention for Short-Term Precipitation Forecasting [9.177158814568887]
Short-term precipitation forecasting model based on ontemporal alignment, with SATA as the temporal alignment module, STAU as the temporal alignment feature extractor.
Based on satellite and ERA5 data, our model achieves improvements of 12.61% in terms of RMSE, in comparison with the state-of-the-art methods.
arXiv Detail & Related papers (2024-09-06T10:28:52Z) - When Rigidity Hurts: Soft Consistency Regularization for Probabilistic
Hierarchical Time Series Forecasting [69.30930115236228]
Probabilistic hierarchical time-series forecasting is an important variant of time-series forecasting.
Most methods focus on point predictions and do not provide well-calibrated probabilistic forecasts distributions.
We propose PROFHiT, a fully probabilistic hierarchical forecasting model that jointly models forecast distribution of entire hierarchy.
arXiv Detail & Related papers (2023-10-17T20:30:16Z) - Streaming Motion Forecasting for Autonomous Driving [71.7468645504988]
We introduce a benchmark that queries future trajectories on streaming data and we refer to it as "streaming forecasting"
Our benchmark inherently captures the disappearance and re-appearance of agents, which is a safety-critical problem yet overlooked by snapshot-based benchmarks.
We propose a plug-and-play meta-algorithm called "Predictive Streamer" that can adapt any snapshot-based forecaster into a streaming forecaster.
arXiv Detail & Related papers (2023-10-02T17:13:16Z) - When Rigidity Hurts: Soft Consistency Regularization for Probabilistic
Hierarchical Time Series Forecasting [69.30930115236228]
Probabilistic hierarchical time-series forecasting is an important variant of time-series forecasting.
Most methods focus on point predictions and do not provide well-calibrated probabilistic forecasts distributions.
We propose PROFHiT, a fully probabilistic hierarchical forecasting model that jointly models forecast distribution of entire hierarchy.
arXiv Detail & Related papers (2022-06-16T06:13:53Z) - Confidence Adaptive Anytime Pixel-Level Recognition [86.75784498879354]
Anytime inference requires a model to make a progression of predictions which might be halted at any time.
We propose the first unified and end-to-end model approach for anytime pixel-level recognition.
arXiv Detail & Related papers (2021-04-01T20:01:57Z) - Long-Short Term Spatiotemporal Tensor Prediction for Passenger Flow
Profile [15.875569404476495]
We focus on a tensor-based prediction and propose several practical techniques to improve prediction.
For long-term prediction specifically, we propose the "Tensor Decomposition + 2-Dimensional Auto-Regressive Moving Average (2D-ARMA)" model.
For short-term prediction, we propose to conduct tensor completion based on tensor clustering to avoid oversimplifying and ensure accuracy.
arXiv Detail & Related papers (2020-04-23T08:30:00Z) - Transformer Hawkes Process [79.16290557505211]
We propose a Transformer Hawkes Process (THP) model, which leverages the self-attention mechanism to capture long-term dependencies.
THP outperforms existing models in terms of both likelihood and event prediction accuracy by a notable margin.
We provide a concrete example, where THP achieves improved prediction performance for learning multiple point processes when incorporating their relational information.
arXiv Detail & Related papers (2020-02-21T13:48:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.