Transfer learning to improve streamflow forecasts in data sparse regions
- URL: http://arxiv.org/abs/2112.03088v1
- Date: Mon, 6 Dec 2021 14:52:53 GMT
- Title: Transfer learning to improve streamflow forecasts in data sparse regions
- Authors: Roland Oruche, Lisa Egede, Tracy Baker, Fearghal O'Donncha
- Abstract summary: We study the methodology behind Transfer Learning (TL) through fine-tuning and parameter transferring for better generalization performance of streamflow prediction in data-sparse regions.
We propose a standard recurrent neural network in the form of Long Short-Term Memory (LSTM) to fit on a sufficiently large source domain dataset.
We present a methodology to implement transfer learning approaches for hydrologic applications by separating the spatial and temporal components of the model and training the model to generalize.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Effective water resource management requires information on water
availability, both in terms of quality and quantity, spatially and temporally.
In this paper, we study the methodology behind Transfer Learning (TL) through
fine-tuning and parameter transferring for better generalization performance of
streamflow prediction in data-sparse regions. We propose a standard recurrent
neural network in the form of Long Short-Term Memory (LSTM) to fit on a
sufficiently large source domain dataset and repurpose the learned weights to a
significantly smaller, yet similar target domain datasets. We present a
methodology to implement transfer learning approaches for spatiotemporal
applications by separating the spatial and temporal components of the model and
training the model to generalize based on categorical datasets representing
spatial variability. The framework is developed on a rich benchmark dataset
from the US and evaluated on a smaller dataset collected by The Nature
Conservancy in Kenya. The LSTM model exhibits generalization performance
through our TL technique. Results from this current experiment demonstrate the
effective predictive skill of forecasting streamflow responses when knowledge
transferring and static descriptors are used to improve hydrologic model
generalization in data-sparse regions.
Related papers
- Hierarchically Disentangled Recurrent Network for Factorizing System Dynamics of Multi-scale Systems [4.634606500665259]
We present a knowledge-guided machine learning (KGML) framework for modeling multi-scale processes.
We study its performance in the context of streamflow forecasting in hydrology.
arXiv Detail & Related papers (2024-07-29T16:25:43Z) - Learn from the Learnt: Source-Free Active Domain Adaptation via Contrastive Sampling and Visual Persistence [60.37934652213881]
Domain Adaptation (DA) facilitates knowledge transfer from a source domain to a related target domain.
This paper investigates a practical DA paradigm, namely Source data-Free Active Domain Adaptation (SFADA), where source data becomes inaccessible during adaptation.
We present learn from the learnt (LFTL), a novel paradigm for SFADA to leverage the learnt knowledge from the source pretrained model and actively iterated models without extra overhead.
arXiv Detail & Related papers (2024-07-26T17:51:58Z) - TPLLM: A Traffic Prediction Framework Based on Pretrained Large Language Models [27.306180426294784]
We introduce TPLLM, a novel traffic prediction framework leveraging Large Language Models (LLMs)
In this framework, we construct a sequence embedding layer based on Conal Neural Networks (LoCNNs) and a graph embedding layer based on Graph Contemporalal Networks (GCNs) to extract sequence features and spatial features.
Experiments on two real-world datasets demonstrate commendable performance in both full-sample and few-shot prediction scenarios.
arXiv Detail & Related papers (2024-03-04T17:08:57Z) - Federated Learning with Projected Trajectory Regularization [65.6266768678291]
Federated learning enables joint training of machine learning models from distributed clients without sharing their local data.
One key challenge in federated learning is to handle non-identically distributed data across the clients.
We propose a novel federated learning framework with projected trajectory regularization (FedPTR) for tackling the data issue.
arXiv Detail & Related papers (2023-12-22T02:12:08Z) - Diffusion Generative Flow Samplers: Improving learning signals through
partial trajectory optimization [87.21285093582446]
Diffusion Generative Flow Samplers (DGFS) is a sampling-based framework where the learning process can be tractably broken down into short partial trajectory segments.
Our method takes inspiration from the theory developed for generative flow networks (GFlowNets)
arXiv Detail & Related papers (2023-10-04T09:39:05Z) - Improving the Transferability of Time Series Forecasting with
Decomposition Adaptation [14.09967794482993]
In time series forecasting, it is difficult to obtain enough data, which limits the performance of neural forecasting models.
To alleviate the data scarcity limitation, we design Sequence Decomposition Adaptation Network (SeDAN)
SeDAN is a novel transfer architecture to improve forecasting performance on the target domain by aligning transferable knowledge from cross-domain datasets.
arXiv Detail & Related papers (2023-06-30T18:12:22Z) - Iterative self-transfer learning: A general methodology for response
time-history prediction based on small dataset [0.0]
An iterative self-transfer learningmethod for training neural networks based on small datasets is proposed in this study.
The results show that the proposed method can improve the model performance by near an order of magnitude on small datasets.
arXiv Detail & Related papers (2023-06-14T18:48:04Z) - Beyond Transfer Learning: Co-finetuning for Action Localisation [64.07196901012153]
We propose co-finetuning -- simultaneously training a single model on multiple upstream'' and downstream'' tasks.
We demonstrate that co-finetuning outperforms traditional transfer learning when using the same total amount of data.
We also show how we can easily extend our approach to multiple upstream'' datasets to further improve performance.
arXiv Detail & Related papers (2022-07-08T10:25:47Z) - How Well Do Sparse Imagenet Models Transfer? [75.98123173154605]
Transfer learning is a classic paradigm by which models pretrained on large "upstream" datasets are adapted to yield good results on "downstream" datasets.
In this work, we perform an in-depth investigation of this phenomenon in the context of convolutional neural networks (CNNs) trained on the ImageNet dataset.
We show that sparse models can match or even outperform the transfer performance of dense models, even at high sparsities.
arXiv Detail & Related papers (2021-11-26T11:58:51Z) - PredRNN: A Recurrent Neural Network for Spatiotemporal Predictive
Learning [109.84770951839289]
We present PredRNN, a new recurrent network for learning visual dynamics from historical context.
We show that our approach obtains highly competitive results on three standard datasets.
arXiv Detail & Related papers (2021-03-17T08:28:30Z) - The Utility of Feature Reuse: Transfer Learning in Data-Starved Regimes [6.419457653976053]
We describe a transfer learning use case for a domain with a data-starved regime.
We evaluate the effectiveness of convolutional feature extraction and fine-tuning.
We conclude that transfer learning enhances the performance of CNN architectures in data-starved regimes.
arXiv Detail & Related papers (2020-02-29T18:48:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.