Precipitation Nowcasting Using Diffusion Transformer with Causal Attention
- URL: http://arxiv.org/abs/2410.13314v1
- Date: Thu, 17 Oct 2024 08:10:41 GMT
- Title: Precipitation Nowcasting Using Diffusion Transformer with Causal Attention
- Authors: ChaoRong Li, XuDong Ling, YiLan Xue, Wenjie Luo, LiHong Zhu, FengQing Qin, Yaodong Zhou, Yuanyuan Huang,
- Abstract summary: Current deep learning methods fall short in establishing effective dependencies between conditions and forecast results.
We propose a Precipitationcasting Using Diffusion Transformer with Causal Attention model.
- Score: 3.9468501770612576
- License:
- Abstract: Short-term precipitation forecasting remains challenging due to the difficulty in capturing long-term spatiotemporal dependencies. Current deep learning methods fall short in establishing effective dependencies between conditions and forecast results, while also lacking interpretability. To address this issue, we propose a Precipitation Nowcasting Using Diffusion Transformer with Causal Attention model. Our model leverages Transformer and combines causal attention mechanisms to establish spatiotemporal queries between conditional information (causes) and forecast results (results). This design enables the model to effectively capture long-term dependencies, allowing forecast results to maintain strong causal relationships with input conditions over a wide range of time and space. We explore four variants of spatiotemporal information interactions for DTCA, demonstrating that global spatiotemporal labeling interactions yield the best performance. In addition, we introduce a Channel-To-Batch shift operation to further enhance the model's ability to represent complex rainfall dynamics. We conducted experiments on two datasets. Compared to state-of-the-art U-Net-based methods, our approach improved the CSI (Critical Success Index) for predicting heavy precipitation by approximately 15% and 8% respectively, achieving state-of-the-art performance.
Related papers
- STAA: Spatio-Temporal Alignment Attention for Short-Term Precipitation Forecasting [9.177158814568887]
Short-term precipitation forecasting model based on ontemporal alignment, with SATA as the temporal alignment module, STAU as the temporal alignment feature extractor.
Based on satellite and ERA5 data, our model achieves improvements of 12.61% in terms of RMSE, in comparison with the state-of-the-art methods.
arXiv Detail & Related papers (2024-09-06T10:28:52Z) - SFANet: Spatial-Frequency Attention Network for Weather Forecasting [54.470205739015434]
Weather forecasting plays a critical role in various sectors, driving decision-making and risk management.
Traditional methods often struggle to capture the complex dynamics of meteorological systems.
We propose a novel framework designed to address these challenges and enhance the accuracy of weather prediction.
arXiv Detail & Related papers (2024-05-29T08:00:15Z) - Efficient Subseasonal Weather Forecast using Teleconnection-informed
Transformers [29.33938664834226]
Subseasonal forecasting is pivotal for agriculture, water resource management, and early warning of disasters.
Recent advances in machine learning have revolutionized weather forecasting by achieving competitive predictive skills to numerical models.
However, training such foundation models requires thousands of GPU days, which causes substantial carbon emissions.
arXiv Detail & Related papers (2024-01-31T14:27:35Z) - GD-CAF: Graph Dual-stream Convolutional Attention Fusion for
Precipitation Nowcasting [1.642094639107215]
We introduce Graph Dual-streamtemporal Conal Attention Fusion (GD-CAF) to learn from historical graph of precipitation maps and nowcast future time step ahead.
GD-CAF consists of gated-temporal convolutional attention as well as fusion modules equipped with depthwise-separable convolutional operations.
We evaluate our model seven years of precipitation maps across Europe and its neighboring areas collected from the ERA5 dataset.
arXiv Detail & Related papers (2024-01-15T20:54:20Z) - Learning from Polar Representation: An Extreme-Adaptive Model for
Long-Term Time Series Forecasting [10.892801642895904]
We propose Distance-weighted Auto-regularized Neural network (DAN), a novel extreme-adaptive model for long-range forecasting of stremflow enhanced by polar representation learning.
On four real-life hydrologic streamflow datasets, we demonstrate that DAN significantly outperforms both state-of-the-art hydrologic time series prediction methods and general methods designed for long-term time series prediction.
arXiv Detail & Related papers (2023-12-14T09:16:01Z) - Learning Robust Precipitation Forecaster by Temporal Frame Interpolation [65.5045412005064]
We develop a robust precipitation forecasting model that demonstrates resilience against spatial-temporal discrepancies.
Our approach has led to significant improvements in forecasting precision, culminating in our model securing textit1st place in the transfer learning leaderboard of the textitWeather4cast'23 competition.
arXiv Detail & Related papers (2023-11-30T08:22:08Z) - Long-term drought prediction using deep neural networks based on geospatial weather data [75.38539438000072]
High-quality drought forecasting up to a year in advance is critical for agriculture planning and insurance.
We tackle drought data by introducing an end-to-end approach that adopts a systematic end-to-end approach.
Key findings are the exceptional performance of a Transformer model, EarthFormer, in making accurate short-term (up to six months) forecasts.
arXiv Detail & Related papers (2023-09-12T13:28:06Z) - The Capacity and Robustness Trade-off: Revisiting the Channel
Independent Strategy for Multivariate Time Series Forecasting [50.48888534815361]
We show that models trained with the Channel Independent (CI) strategy outperform those trained with the Channel Dependent (CD) strategy.
Our results conclude that the CD approach has higher capacity but often lacks robustness to accurately predict distributionally drifted time series.
We propose a modified CD method called Predict Residuals with Regularization (PRReg) that can surpass the CI strategy.
arXiv Detail & Related papers (2023-04-11T13:15:33Z) - Towards Long-Term Time-Series Forecasting: Feature, Pattern, and
Distribution [57.71199089609161]
Long-term time-series forecasting (LTTF) has become a pressing demand in many applications, such as wind power supply planning.
Transformer models have been adopted to deliver high prediction capacity because of the high computational self-attention mechanism.
We propose an efficient Transformerbased model, named Conformer, which differentiates itself from existing methods for LTTF in three aspects.
arXiv Detail & Related papers (2023-01-05T13:59:29Z) - TRU-NET: A Deep Learning Approach to High Resolution Prediction of
Rainfall [21.399707529966474]
We present TRU-NET, an encoder-decoder model featuring a novel 2D cross attention mechanism between contiguous convolutional-recurrent layers.
We use a conditional-continuous loss function to capture the zero-skewed %extreme event patterns of rainfall.
Experiments show that our model consistently attains lower RMSE and MAE scores than a DL model prevalent in short term precipitation prediction.
arXiv Detail & Related papers (2020-08-20T17:27:59Z) - Transformer Hawkes Process [79.16290557505211]
We propose a Transformer Hawkes Process (THP) model, which leverages the self-attention mechanism to capture long-term dependencies.
THP outperforms existing models in terms of both likelihood and event prediction accuracy by a notable margin.
We provide a concrete example, where THP achieves improved prediction performance for learning multiple point processes when incorporating their relational information.
arXiv Detail & Related papers (2020-02-21T13:48:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.