Continental-scale streamflow modeling of basins with reservoirs: a
demonstration of effectiveness and a delineation of challenges
- URL: http://arxiv.org/abs/2101.04423v1
- Date: Tue, 12 Jan 2021 11:49:54 GMT
- Title: Continental-scale streamflow modeling of basins with reservoirs: a
demonstration of effectiveness and a delineation of challenges
- Authors: Wenyu Ouyang, Kathryn Lawson, Dapeng Feng, Lei Ye, Chi Zhang, Chaopeng
Shen
- Abstract summary: A large fraction of major waterways have dams influencing streamflow, which must be accounted for in large-scale hydrologic modeling.
Here we take a divide-and-conquer approach to examine which types of basins could be well represented by a long short-term memory (LSTM) deep learning model.
We analyzed data from 3557 basins (83% dammed) over the contiguous United States and noted strong impacts of reservoir purposes, capacity-to-runoff ratio (dor), and diversion on streamflow.
- Score: 4.834945446235863
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A large fraction of major waterways have dams influencing streamflow, which
must be accounted for in large-scale hydrologic modeling. However, daily
streamflow prediction for basins with dams is challenging for various modeling
approaches, especially at large scales. Here we took a divide-and-conquer
approach to examine which types of basins could be well represented by a long
short-term memory (LSTM) deep learning model using only readily-available
information. We analyzed data from 3557 basins (83% dammed) over the contiguous
United States and noted strong impacts of reservoir purposes,
capacity-to-runoff ratio (dor), and diversion on streamflow on streamflow
modeling. Surprisingly, while the LSTM model trained on a widely-used
reference-basin dataset performed poorly for more non-reference basins, the
model trained on the whole dataset presented a median test Nash-Sutcliffe
efficiency coefficient (NSE) of 0.74, reaching benchmark-level performance. The
zero-dor, small-dor, and large-dor basins were found to have distinct
behaviors, so migrating models between categories yielded catastrophic results.
However, training with pooled data from different sets yielded optimal median
NSEs of 0.73, 0.78, and 0.71 for these groups, respectively, showing noticeable
advantages over existing models. These results support a coherent, mixed
modeling strategy where smaller dams are modeled as part of rainfall-runoff
processes, but dammed basins must not be treated as reference ones and must be
included in the training set; then, large-dor reservoirs can be represented
explicitly and future work should examine modeling reservoirs for fire
protection and irrigation, followed by those for hydroelectric power
generation, and flood control, etc.
Related papers
- Evaluating Deep Learning Approaches for Predictions in Unmonitored Basins with Continental-scale Stream Temperature Models [1.8067095934521364]
Recent machine learning (ML) models can harness vast datasets for accurate predictions at large spatial scales.
This study explores questions regarding model design and data needed for inputs and training to improve performance.
arXiv Detail & Related papers (2024-10-23T15:36:59Z) - Hierarchically Disentangled Recurrent Network for Factorizing System Dynamics of Multi-scale Systems [4.634606500665259]
We present a knowledge-guided machine learning (KGML) framework for modeling multi-scale processes.
We study its performance in the context of streamflow forecasting in hydrology.
arXiv Detail & Related papers (2024-07-29T16:25:43Z) - No "Zero-Shot" Without Exponential Data: Pretraining Concept Frequency Determines Multimodal Model Performance [68.18779562801762]
multimodal models require exponentially more data to achieve linear improvements in downstream "zero-shot" performance.
Our study reveals an exponential need for training data which implies that the key to "zero-shot" generalization capabilities under large-scale training paradigms remains to be found.
arXiv Detail & Related papers (2024-04-04T17:58:02Z) - The Languini Kitchen: Enabling Language Modelling Research at Different
Scales of Compute [66.84421705029624]
We introduce an experimental protocol that enables model comparisons based on equivalent compute, measured in accelerator hours.
We pre-process an existing large, diverse, and high-quality dataset of books that surpasses existing academic benchmarks in quality, diversity, and document length.
This work also provides two baseline models: a feed-forward model derived from the GPT-2 architecture and a recurrent model in the form of a novel LSTM with ten-fold throughput.
arXiv Detail & Related papers (2023-09-20T10:31:17Z) - Learning Large-scale Subsurface Simulations with a Hybrid Graph Network
Simulator [57.57321628587564]
We introduce Hybrid Graph Network Simulator (HGNS) for learning reservoir simulations of 3D subsurface fluid flows.
HGNS consists of a subsurface graph neural network (SGNN) to model the evolution of fluid flows, and a 3D-U-Net to model the evolution of pressure.
Using an industry-standard subsurface flow dataset (SPE-10) with 1.1 million cells, we demonstrate that HGNS is able to reduce the inference time up to 18 times compared to standard subsurface simulators.
arXiv Detail & Related papers (2022-06-15T17:29:57Z) - Deep Equilibrium Optical Flow Estimation [80.80992684796566]
Recent state-of-the-art (SOTA) optical flow models use finite-step recurrent update operations to emulate traditional algorithms.
These RNNs impose large computation and memory overheads, and are not directly trained to model such stable estimation.
We propose deep equilibrium (DEQ) flow estimators, an approach that directly solves for the flow as the infinite-level fixed point of an implicit layer.
arXiv Detail & Related papers (2022-04-18T17:53:44Z) - Differentiable, learnable, regionalized process-based models with
physical outputs can approach state-of-the-art hydrologic prediction accuracy [1.181206257787103]
We show that differentiable, learnable, process-based models (called delta models here) can approach the performance level of LSTM for the intensively-observed variable (streamflow) with regionalized parameterization.
We use a simple hydrologic model HBV as the backbone and use embedded neural networks, which can only be trained in a differentiable programming framework.
arXiv Detail & Related papers (2022-03-28T15:06:53Z) - Churn Reduction via Distillation [54.5952282395487]
We show an equivalence between training with distillation using the base model as the teacher and training with an explicit constraint on the predictive churn.
We then show that distillation performs strongly for low churn training against a number of recent baselines.
arXiv Detail & Related papers (2021-06-04T18:03:31Z) - High Temporal Resolution Rainfall Runoff Modelling Using
Long-Short-Term-Memory (LSTM) Networks [0.03694429692322631]
The model was tested for a watershed in Houston, TX, known for severe flood events.
The LSTM network's capability in learning long-term dependencies between the input and output of the network allowed modeling RR with high resolution in time.
arXiv Detail & Related papers (2020-02-07T00:38:03Z) - Model Reuse with Reduced Kernel Mean Embedding Specification [70.044322798187]
We present a two-phase framework for finding helpful models for a current application.
In the upload phase, when a model is uploading into the pool, we construct a reduced kernel mean embedding (RKME) as a specification for the model.
Then in the deployment phase, the relatedness of the current task and pre-trained models will be measured based on the value of the RKME specification.
arXiv Detail & Related papers (2020-01-20T15:15:07Z) - Stream-Flow Forecasting of Small Rivers Based on LSTM [3.921808417990452]
This paper tries to provide a new method to do the forecast using the Long-Short Term Memory (LSTM) deep learning model.
We collected the stream flow data from one hydrologic station in Tunxi, China, and precipitation data from 11 rainfall stations around to forecast the stream flow data.
We evaluated the prediction results using three criteria: root mean square error (RMSE), mean absolute error (MAE), and coefficient of determination (R2)
arXiv Detail & Related papers (2020-01-16T07:14:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.