A Hybrid Deep-Learning Model for El Niño Southern Oscillation in the Low-Data Regime
- URL: http://arxiv.org/abs/2412.03743v2
- Date: Sun, 27 Apr 2025 12:58:08 GMT
- Title: A Hybrid Deep-Learning Model for El Niño Southern Oscillation in the Low-Data Regime
- Authors: Jakob Schloer, Matthew Newman, Jannik Thuemmel, Antonietta Capotondi, Bedartha Goswami,
- Abstract summary: El Nino Southern Oscillation (ENSO) forecasts can be made up to one year in advance.<n>Deep-learning models are predominantly trained on climate model simulations that provide thousands of years of training data.<n>This motivates a hybrid approach, combining the LIMs modest data needs with a deep-learning non-Markovian correction of the LIM.<n>For O(100 yr) datasets, our resulting Hybrid model is more skillful than the LIM while also exceeding the skill of a full deep-learning model.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: While deep-learning models have demonstrated skillful El Ni\~no Southern Oscillation (ENSO) forecasts up to one year in advance, they are predominantly trained on climate model simulations that provide thousands of years of training data at the expense of introducing climate model biases. Simpler Linear Inverse Models (LIMs) trained on the much shorter observational record also make skillful ENSO predictions but do not capture predictable nonlinear processes. This motivates a hybrid approach, combining the LIMs modest data needs with a deep-learning non-Markovian correction of the LIM. For O(100 yr) datasets, our resulting Hybrid model is more skillful than the LIM while also exceeding the skill of a full deep-learning model. Additionally, while the most predictable ENSO events are still identified in advance by the LIM, they are better predicted by the Hybrid model, especially in the western tropical Pacific for leads beyond about 9 months, by capturing the subsequent asymmetric (warm versus cold phases) evolution of ENSO.
Related papers
- Long-Range Distillation: Distilling 10,000 Years of Simulated Climate into Long Timestep AI Weather Models [0.0]
Long-range distillation is a method that trains a long-timestep probabilistic "student" model to forecast directly at long-range.<n>We generate over 10,000 years of simulated climate to train models for forecasting across a range of timescales.<n>In perfect-model experiments, the distilled models outperform climatology and approach the skill of their autoregressive teacher.
arXiv Detail & Related papers (2025-12-28T07:03:20Z) - Exploring Machine Learning, Deep Learning, and Explainable AI Methods for Seasonal Precipitation Prediction in South America [2.084547583236646]
This study investigates the feasibility of purely data-driven approaches for precipitation forecasting in South America.<n>The selected classical ML techniques were Random Forests and extreme gradient boosting (XGBoost), while the DL counterparts were a 1D convolutional neural network (CNN 1D), a long short-term memory (LSTM) model, and a gated recurrent unit (GRU) model.<n>The results of this research confirm the viability of DL models for climate forecasting, solidifying a global trend in major meteorological and climate forecasting centers.
arXiv Detail & Related papers (2025-12-15T21:37:27Z) - When Simpler Wins: Facebooks Prophet vs LSTM for Air Pollution Forecasting in Data-Constrained Northern Nigeria [0.44198435146063364]
This study evaluates Long Short-Term Memory (LSTM) networks and the Facebook Prophet model for forecasting multiple pollutants.<n>Results show that Prophet often matches or exceeds LSTM's accuracy, particularly in series dominated by seasonal and long-term trends.
arXiv Detail & Related papers (2025-08-22T09:23:59Z) - On conditional diffusion models for PDE simulations [53.01911265639582]
We study score-based diffusion models for forecasting and assimilation of sparse observations.
We propose an autoregressive sampling approach that significantly improves performance in forecasting.
We also propose a new training strategy for conditional score-based models that achieves stable performance over a range of history lengths.
arXiv Detail & Related papers (2024-10-21T18:31:04Z) - Embedding machine-learnt sub-grid variability improves climate model biases [0.44998333629984877]
Under-representation of cloud formation is a long-standing bias associated with climate simulations.
We overcome these biases by embedding a Multi-Output Gaussian Process (MOGP) trained on high resolution Unified Model simulations.
Ten-year predictions are generated for both control and ML-hybrid models.
arXiv Detail & Related papers (2024-06-13T19:35:58Z) - Weather Prediction with Diffusion Guided by Realistic Forecast Processes [49.07556359513563]
We introduce a novel method that applies diffusion models (DM) for weather forecasting.
Our method can achieve both direct and iterative forecasting with the same modeling framework.
The flexibility and controllability of our model empowers a more trustworthy DL system for the general weather community.
arXiv Detail & Related papers (2024-02-06T21:28:42Z) - ExtremeCast: Boosting Extreme Value Prediction for Global Weather Forecast [57.6987191099507]
We introduce Exloss, a novel loss function that performs asymmetric optimization and highlights extreme values to obtain accurate extreme weather forecast.
We also introduce ExBooster, which captures the uncertainty in prediction outcomes by employing multiple random samples.
Our solution can achieve state-of-the-art performance in extreme weather prediction, while maintaining the overall forecast accuracy comparable to the top medium-range forecast models.
arXiv Detail & Related papers (2024-02-02T10:34:13Z) - Long-term drought prediction using deep neural networks based on geospatial weather data [75.38539438000072]
High-quality drought forecasting up to a year in advance is critical for agriculture planning and insurance.
We tackle drought data by introducing an end-to-end approach that adopts a systematic end-to-end approach.
Key findings are the exceptional performance of a Transformer model, EarthFormer, in making accurate short-term (up to six months) forecasts.
arXiv Detail & Related papers (2023-09-12T13:28:06Z) - Learning Sample Difficulty from Pre-trained Models for Reliable
Prediction [55.77136037458667]
We propose to utilize large-scale pre-trained models to guide downstream model training with sample difficulty-aware entropy regularization.
We simultaneously improve accuracy and uncertainty calibration across challenging benchmarks.
arXiv Detail & Related papers (2023-04-20T07:29:23Z) - Stabilizing Machine Learning Prediction of Dynamics: Noise and
Noise-inspired Regularization [58.720142291102135]
Recent has shown that machine learning (ML) models can be trained to accurately forecast the dynamics of chaotic dynamical systems.
In the absence of mitigating techniques, this technique can result in artificially rapid error growth, leading to inaccurate predictions and/or climate instability.
We introduce Linearized Multi-Noise Training (LMNT), a regularization technique that deterministically approximates the effect of many small, independent noise realizations added to the model input during training.
arXiv Detail & Related papers (2022-11-09T23:40:52Z) - Surrogate Ensemble Forecasting for Dynamic Climate Impact Models [0.0]
This study considers a climate driven disease model, the Liverpool Malaria Model (LMM), which predicts the malaria transmission coefficient R0.
The input and output data is used to train surrogate models in the form of a Random Forest Quantile Regression (RFQR) model and a Bayesian Long Short-Term Memory (BLSTM) neural network.
arXiv Detail & Related papers (2022-04-12T13:30:01Z) - Learning and Dynamical Models for Sub-seasonal Climate Forecasting:
Comparison and Collaboration [20.52175766498954]
Sub-seasonal climate forecasting (SSF) is the prediction of key climate variables such as temperature and precipitation on the 2-week to 2-month time horizon.
Recent studies have shown the potential of machine learning (ML) models to advance SSF.
arXiv Detail & Related papers (2021-09-29T06:34:34Z) - Hybrid Physics and Deep Learning Model for Interpretable Vehicle State
Prediction [75.1213178617367]
We propose a hybrid approach combining deep learning and physical motion models.
We achieve interpretability by restricting the output range of the deep neural network as part of the hybrid model.
The results show that our hybrid model can improve model interpretability with no decrease in accuracy compared to existing deep learning approaches.
arXiv Detail & Related papers (2021-03-11T15:21:08Z) - Reservoir Computing as a Tool for Climate Predictability Studies [0.0]
We show that Reservoir Computing provides an alternative nonlinear approach that improves on the predictive skill of the Linear-Inverse-Modeling approach.
The improved predictive skill of the RC approach over a wide range of conditions suggests that this machine-learning technique may have a use in climate predictability studies.
arXiv Detail & Related papers (2021-02-24T22:22:59Z) - Short-term daily precipitation forecasting with seasonally-integrated
autoencoder [0.0]
We propose a seasonally-integrated autoencoder (SSAE) consisting of two long short-term memory (LSTM) autoencoders.
Our experimental results show that not only does the SSAE outperform various time series models regardless of the climate type, but it also has low output variance compared to other deep learning models.
arXiv Detail & Related papers (2021-01-23T14:19:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.