Forecasting Corn Yield with Machine Learning Ensembles
- URL: http://arxiv.org/abs/2001.09055v2
- Date: Fri, 6 Nov 2020 18:20:59 GMT
- Title: Forecasting Corn Yield with Machine Learning Ensembles
- Authors: Mohsen Shahhosseini, Guiping Hu, Sotirios V. Archontoulis
- Abstract summary: This paper provides a machine learning based framework to forecast corn yields in three US Corn Belt states (Illinois, Indiana, and Iowa)
Several ensemble models are designed using blocked sequential procedure to generate out-of-bag predictions.
Results show that ensemble models based on weighted average of the base learners outperform individual models.
- Score: 2.9005223064604078
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The emerge of new technologies to synthesize and analyze big data with
high-performance computing, has increased our capacity to more accurately
predict crop yields. Recent research has shown that Machine learning (ML) can
provide reasonable predictions, faster, and with higher flexibility compared to
simulation crop modeling. The earlier the prediction during the growing season
the better, but this has not been thoroughly investigated as previous studies
considered all data available to predict yields. This paper provides a machine
learning based framework to forecast corn yields in three US Corn Belt states
(Illinois, Indiana, and Iowa) considering complete and partial in-season
weather knowledge. Several ensemble models are designed using blocked
sequential procedure to generate out-of-bag predictions. The forecasts are made
in county-level scale and aggregated for agricultural district, and state level
scales. Results show that ensemble models based on weighted average of the base
learners outperform individual models. Specifically, the proposed ensemble
model could achieve best prediction accuracy (RRMSE of 7.8%) and least mean
bias error (-6.06 bu/acre) compared to other developed models. Comparing our
proposed model forecasts with the literature demonstrates the superiority of
forecasts made by our proposed ensemble model. Results from the scenario of
having partial in-season weather knowledge reveal that decent yield forecasts
can be made as early as June 1st. To find the marginal effect of each input
feature on the forecasts made by the proposed ensemble model, a methodology is
suggested that is the basis for finding feature importance for the ensemble
model. The findings suggest that weather features corresponding to weather in
weeks 18-24 (May 1st to June 1st) are the most important input features.
Related papers
- Forecasting with Deep Learning: Beyond Average of Average of Average Performance [0.393259574660092]
Current practices for evaluating and comparing forecasting models focus on summarising performance into a single score.
We propose a novel framework for evaluating models from multiple perspectives.
We show the advantages of this framework by comparing a state-of-the-art deep learning approach with classical forecasting techniques.
arXiv Detail & Related papers (2024-06-24T12:28:22Z) - Generative weather for improved crop model simulations [0.0]
We propose a new method to construct generative models for long-term weather forecasts.
Results show significant improvement from the conventional method.
For individual crop modellers to start applying the method to their problems, technical details are carefully explained.
arXiv Detail & Related papers (2024-03-31T02:03:28Z) - Uncertainty quantification for data-driven weather models [0.0]
We study and compare uncertainty quantification methods to generate probabilistic weather forecasts from a state-of-the-art deterministic data-driven weather model, Pangu-Weather.
Specifically, we compare approaches for quantifying forecast uncertainty based on generating ensemble forecasts via perturbations to the initial conditions.
In a case study on medium-range forecasts of selected weather variables over Europe, the probabilistic forecasts obtained by using the Pangu-Weather model in concert with uncertainty quantification methods show promising results.
arXiv Detail & Related papers (2024-03-20T10:07:51Z) - Predictive Churn with the Set of Good Models [64.05949860750235]
We study the effect of conflicting predictions over the set of near-optimal machine learning models.
We present theoretical results on the expected churn between models within the Rashomon set.
We show how our approach can be used to better anticipate, reduce, and avoid churn in consumer-facing applications.
arXiv Detail & Related papers (2024-02-12T16:15:25Z) - ExtremeCast: Boosting Extreme Value Prediction for Global Weather Forecast [57.6987191099507]
We introduce Exloss, a novel loss function that performs asymmetric optimization and highlights extreme values to obtain accurate extreme weather forecast.
We also introduce ExBooster, which captures the uncertainty in prediction outcomes by employing multiple random samples.
Our solution can achieve state-of-the-art performance in extreme weather prediction, while maintaining the overall forecast accuracy comparable to the top medium-range forecast models.
arXiv Detail & Related papers (2024-02-02T10:34:13Z) - Attention-Based Ensemble Pooling for Time Series Forecasting [55.2480439325792]
We propose a method for pooling that performs a weighted average over candidate model forecasts.
We test this method on two time-series forecasting problems: multi-step forecasting of the dynamics of the non-stationary Lorenz 63 equation, and one-step forecasting of the weekly incident deaths due to COVID-19.
arXiv Detail & Related papers (2023-10-24T22:59:56Z) - Long-term drought prediction using deep neural networks based on geospatial weather data [75.38539438000072]
High-quality drought forecasting up to a year in advance is critical for agriculture planning and insurance.
We tackle drought data by introducing an end-to-end approach that adopts a systematic end-to-end approach.
Key findings are the exceptional performance of a Transformer model, EarthFormer, in making accurate short-term (up to six months) forecasts.
arXiv Detail & Related papers (2023-09-12T13:28:06Z) - Beyond S-curves: Recurrent Neural Networks for Technology Forecasting [60.82125150951035]
We develop an autencoder approach that employs recent advances in machine learning and time series forecasting.
S-curves forecasts largely exhibit a mean average percentage error (MAPE) comparable to a simple ARIMA baseline.
Our autoencoder approach improves the MAPE by 13.5% on average over the second-best result.
arXiv Detail & Related papers (2022-11-28T14:16:22Z) - Computing the ensemble spread from deterministic weather predictions
using conditional generative adversarial networks [0.0]
We propose to use deep-learning algorithms to learn the statistical properties of an ensemble prediction system.
Once trained, the costly ensemble prediction system will not be needed anymore to obtain future ensemble forecasts.
arXiv Detail & Related papers (2022-05-18T19:10:38Z) - Feature-weighted Stacking for Nonseasonal Time Series Forecasts: A Case
Study of the COVID-19 Epidemic Curves [0.0]
We investigate ensembling techniques in forecasting and examine their potential for use in nonseasonal time-series.
We propose using late data fusion, using a stacked ensemble of two forecasting models and two meta-features that prove their predictive power during a preliminary forecasting stage.
arXiv Detail & Related papers (2021-08-19T14:44:46Z) - Test-time Collective Prediction [73.74982509510961]
Multiple parties in machine learning want to jointly make predictions on future test points.
Agents wish to benefit from the collective expertise of the full set of agents, but may not be willing to release their data or model parameters.
We explore a decentralized mechanism to make collective predictions at test time, leveraging each agent's pre-trained model.
arXiv Detail & Related papers (2021-06-22T18:29:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.