The data synergy effects of time-series deep learning models in
hydrology
- URL: http://arxiv.org/abs/2101.01876v1
- Date: Wed, 6 Jan 2021 05:24:45 GMT
- Title: The data synergy effects of time-series deep learning models in
hydrology
- Authors: Kuai Fang, Daniel Kifer, Kathryn Lawson, Dapeng Feng, Chaopeng Shen
- Abstract summary: We argue that unification can significantly outperform regionalization in the era of big data and deep learning (DL)
We highlight an effect we call data synergy, where the results of the DL models improved when data were pooled together from characteristically different regions.
- Score: 9.282656246381102
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: When fitting statistical models to variables in geoscientific disciplines
such as hydrology, it is a customary practice to regionalize - to divide a
large spatial domain into multiple regions and study each region separately -
instead of fitting a single model on the entire data (also known as
unification). Traditional wisdom in these fields suggests that models built for
each region separately will have higher performance because of homogeneity
within each region. However, by partitioning the training data, each model has
access to fewer data points and cannot learn from commonalities between
regions. Here, through two hydrologic examples (soil moisture and streamflow),
we argue that unification can often significantly outperform regionalization in
the era of big data and deep learning (DL). Common DL architectures, even
without bespoke customization, can automatically build models that benefit from
regional commonality while accurately learning region-specific differences. We
highlight an effect we call data synergy, where the results of the DL models
improved when data were pooled together from characteristically different
regions. In fact, the performance of the DL models benefited from more diverse
rather than more homogeneous training data. We hypothesize that DL models
automatically adjust their internal representations to identify commonalities
while also providing sufficient discriminatory information to the model. The
results here advocate for pooling together larger datasets, and suggest the
academic community should place greater emphasis on data sharing and
compilation.
Related papers
- Learning Divergence Fields for Shift-Robust Graph Representations [73.11818515795761]
In this work, we propose a geometric diffusion model with learnable divergence fields for the challenging problem with interdependent data.
We derive a new learning objective through causal inference, which can guide the model to learn generalizable patterns of interdependence that are insensitive across domains.
arXiv Detail & Related papers (2024-06-07T14:29:21Z) - SARN: Structurally-Aware Recurrent Network for Spatio-Temporal Disaggregation [8.636014676778682]
Open data is frequently released spatially aggregated, usually to comply with privacy policies. But coarse, heterogeneous aggregations complicate coherent learning and integration for downstream AI/ML systems.
We propose an overarching model named Structurally-Aware Recurrent Network (SARN), which integrates structurally-aware spatial attention layers into the Gated Recurrent Unit (GRU) model.
For scenarios with limited historical training data, we show that a model pre-trained on one city variable can be fine-tuned for another city variable using only a few hundred samples.
arXiv Detail & Related papers (2023-06-09T21:01:29Z) - Federated Learning of Models Pre-Trained on Different Features with
Consensus Graphs [19.130197923214123]
Learning an effective global model on private and decentralized datasets has become an increasingly important challenge of machine learning.
We propose a feature fusion approach that extracts local representations from local models and incorporates them into a global representation that improves the prediction performance.
This paper presents solutions to these problems and demonstrates them in real-world applications on time series data such as power grids and traffic networks.
arXiv Detail & Related papers (2023-06-02T02:24:27Z) - SALUDA: Surface-based Automotive Lidar Unsupervised Domain Adaptation [62.889835139583965]
We introduce an unsupervised auxiliary task of learning an implicit underlying surface representation simultaneously on source and target data.
As both domains share the same latent representation, the model is forced to accommodate discrepancies between the two sources of data.
Our experiments demonstrate that our method achieves a better performance than the current state of the art, both in real-to-real and synthetic-to-real scenarios.
arXiv Detail & Related papers (2023-04-06T17:36:23Z) - Knowledge is a Region in Weight Space for Fine-tuned Language Models [48.589822853418404]
We study how the weight space and the underlying loss landscape of different models are interconnected.
We show that language models that have been finetuned on the same dataset form a tight cluster in the weight space, while models finetuned on different datasets from the same underlying task form a looser cluster.
arXiv Detail & Related papers (2023-02-09T18:59:18Z) - FedSiam-DA: Dual-aggregated Federated Learning via Siamese Network under
Non-IID Data [21.95009868875851]
Federated learning can address data island, it remains challenging to train with data heterogeneous in a real application.
We propose FedSiam-DA, a novel dual-aggregated contrastive federated learning approach.
arXiv Detail & Related papers (2022-11-17T09:05:25Z) - Towards Understanding and Mitigating Dimensional Collapse in Heterogeneous Federated Learning [112.69497636932955]
Federated learning aims to train models across different clients without the sharing of data for privacy considerations.
We study how data heterogeneity affects the representations of the globally aggregated models.
We propose sc FedDecorr, a novel method that can effectively mitigate dimensional collapse in federated learning.
arXiv Detail & Related papers (2022-10-01T09:04:17Z) - CHALLENGER: Training with Attribution Maps [63.736435657236505]
We show that utilizing attribution maps for training neural networks can improve regularization of models and thus increase performance.
In particular, we show that our generic domain-independent approach yields state-of-the-art results in vision, natural language processing and on time series tasks.
arXiv Detail & Related papers (2022-05-30T13:34:46Z) - Dataset Cartography: Mapping and Diagnosing Datasets with Training
Dynamics [118.75207687144817]
We introduce Data Maps, a model-based tool to characterize and diagnose datasets.
We leverage a largely ignored source of information: the behavior of the model on individual instances during training.
Our results indicate that a shift in focus from quantity to quality of data could lead to robust models and improved out-of-distribution generalization.
arXiv Detail & Related papers (2020-09-22T20:19:41Z) - Meta-Learning for Few-Shot Land Cover Classification [3.8529010979482123]
We evaluate the model-agnostic meta-learning (MAML) algorithm on classification and segmentation tasks.
We find that few-shot model adaptation outperforms pre-training with regular gradient descent.
This indicates that model optimization with meta-learning may benefit tasks in the Earth sciences.
arXiv Detail & Related papers (2020-04-28T09:42:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.