Cooperative data-driven modeling
- URL: http://arxiv.org/abs/2211.12971v2
- Date: Fri, 8 Mar 2024 16:40:10 GMT
- Title: Cooperative data-driven modeling
- Authors: Aleksandr Dekhovich, O. Taylan Turan, Jiaxiang Yi, Miguel A. Bessa
- Abstract summary: Data-driven modeling in mechanics is evolving rapidly based on recent machine learning advances.
New data and models created by different groups become available, opening possibilities for cooperative modeling.
Artificial neural networks suffer from catastrophic forgetting, i.e. they forget how to perform an old task when trained on a new one.
This hinders cooperation because adapting an existing model for a new task affects the performance on a previous task trained by someone else.
- Score: 44.99833362998488
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Data-driven modeling in mechanics is evolving rapidly based on recent machine
learning advances, especially on artificial neural networks. As the field
matures, new data and models created by different groups become available,
opening possibilities for cooperative modeling. However, artificial neural
networks suffer from catastrophic forgetting, i.e. they forget how to perform
an old task when trained on a new one. This hinders cooperation because
adapting an existing model for a new task affects the performance on a previous
task trained by someone else. The authors developed a continual learning method
that addresses this issue, applying it here for the first time to solid
mechanics. In particular, the method is applied to recurrent neural networks to
predict history-dependent plasticity behavior, although it can be used on any
other architecture (feedforward, convolutional, etc.) and to predict other
phenomena. This work intends to spawn future developments on continual learning
that will foster cooperative strategies among the mechanics community to solve
increasingly challenging problems. We show that the chosen continual learning
strategy can sequentially learn several constitutive laws without forgetting
them, using less data to achieve the same error as standard (non-cooperative)
training of one law per model.
Related papers
- Transferable Post-training via Inverse Value Learning [83.75002867411263]
We propose modeling changes at the logits level during post-training using a separate neural network (i.e., the value network)
After training this network on a small base model using demonstrations, this network can be seamlessly integrated with other pre-trained models during inference.
We demonstrate that the resulting value network has broad transferability across pre-trained models of different parameter sizes.
arXiv Detail & Related papers (2024-10-28T13:48:43Z) - Adapt & Align: Continual Learning with Generative Models Latent Space
Alignment [15.729732755625474]
We introduce Adapt & Align, a method for continual learning of neural networks by aligning latent representations in generative models.
Neural Networks suffer from abrupt loss in performance when retrained with additional data.
We propose a new method that mitigates those problems by employing generative models and splitting the process of their update into two parts.
arXiv Detail & Related papers (2023-12-21T10:02:17Z) - Negotiated Representations to Prevent Forgetting in Machine Learning
Applications [0.0]
Catastrophic forgetting is a significant challenge in the field of machine learning.
We propose a novel method for preventing catastrophic forgetting in machine learning applications.
arXiv Detail & Related papers (2023-11-30T22:43:50Z) - A Survey on Knowledge Editing of Neural Networks [43.813073385305806]
Even the largest artificial neural networks make mistakes, and once-correct predictions can become invalid as the world progresses in time.
Knowledge editing is emerging as a novel area of research that aims to enable reliable, data-efficient, and fast changes to a pre-trained target model.
We first introduce the problem of editing neural networks, formalize it in a common framework and differentiate it from more notorious branches of research such as continuous learning.
arXiv Detail & Related papers (2023-10-30T16:29:47Z) - Learning-based adaption of robotic friction models [48.453527255659296]
We introduce a novel approach to adapt an existing friction model to new dynamics using as little data as possible.
Our proposed estimator outperforms the conventional model-based approach and the base neural network significantly.
Our method does not rely on data with external load during training, eliminating the need for external torque sensors.
arXiv Detail & Related papers (2023-10-25T14:50:15Z) - Continual Learning with Transformers for Image Classification [12.028617058465333]
In computer vision, neural network models struggle to continually learn new concepts without forgetting what has been learnt in the past.
We develop a solution called Adaptive Distillation of Adapters (ADA), which is developed to perform continual learning.
We empirically demonstrate on different classification tasks that this method maintains a good predictive performance without retraining the model.
arXiv Detail & Related papers (2022-06-28T15:30:10Z) - Continual Learning with Bayesian Model based on a Fixed Pre-trained
Feature Extractor [55.9023096444383]
Current deep learning models are characterised by catastrophic forgetting of old knowledge when learning new classes.
Inspired by the process of learning new knowledge in human brains, we propose a Bayesian generative model for continual learning.
arXiv Detail & Related papers (2022-04-28T08:41:51Z) - EINNs: Epidemiologically-Informed Neural Networks [75.34199997857341]
We introduce a new class of physics-informed neural networks-EINN-crafted for epidemic forecasting.
We investigate how to leverage both the theoretical flexibility provided by mechanistic models as well as the data-driven expressability afforded by AI models.
arXiv Detail & Related papers (2022-02-21T18:59:03Z) - Reducing Catastrophic Forgetting in Self Organizing Maps with
Internally-Induced Generative Replay [67.50637511633212]
A lifelong learning agent is able to continually learn from potentially infinite streams of pattern sensory data.
One major historic difficulty in building agents that adapt is that neural systems struggle to retain previously-acquired knowledge when learning from new samples.
This problem is known as catastrophic forgetting (interference) and remains an unsolved problem in the domain of machine learning to this day.
arXiv Detail & Related papers (2021-12-09T07:11:14Z) - Learning to Transfer Dynamic Models of Underactuated Soft Robotic Hands [15.481728234509227]
Transfer learning is a popular approach to bypassing data limitations in one domain by leveraging data from another domain.
We show that in some situations this can lead to significantly worse performance than simply using the transferred model without adaptation.
We derive an upper bound on the Lyapunov exponent of a trained transition model, and demonstrate two approaches that make use of this insight.
arXiv Detail & Related papers (2020-05-21T01:46:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.