Transfer learning for predicting source terms of principal component
transport in chemically reactive flow
- URL: http://arxiv.org/abs/2312.00356v1
- Date: Fri, 1 Dec 2023 05:18:35 GMT
- Title: Transfer learning for predicting source terms of principal component
transport in chemically reactive flow
- Authors: Ki Sung Jung, Tarek Echekki, Jacqueline H. Chen, Mohammad Khalil
- Abstract summary: The aim of this study is to evaluate whether the number of requisite training samples can be reduced with the use of various transfer learning models.
Three transfer learning strategies are then applied to the training of the ANN model with a sparse dataset.
The performance of the reduced-order model with a sparse dataset is found to be remarkably enhanced if the training of the ANN model is restricted by a regularization term.
- Score: 0.40964539027092917
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The objective of this study is to evaluate whether the number of requisite
training samples can be reduced with the use of various transfer learning
models for predicting, for example, the chemical source terms of the
data-driven reduced-order model that represents the homogeneous ignition
process of a hydrogen/air mixture. Principal component analysis is applied to
reduce the dimensionality of the hydrogen/air mixture in composition space.
Artificial neural networks (ANNs) are used to tabulate the reaction rates of
principal components, and subsequently, a system of ordinary differential
equations is solved. As the number of training samples decreases at the target
task (i.e.,for T0 > 1000 K and various phi), the reduced-order model fails to
predict the ignition evolution of a hydrogen/air mixture. Three transfer
learning strategies are then applied to the training of the ANN model with a
sparse dataset. The performance of the reduced-order model with a sparse
dataset is found to be remarkably enhanced if the training of the ANN model is
restricted by a regularization term that controls the degree of knowledge
transfer from source to target tasks. To this end, a novel transfer learning
method is introduced, parameter control via partial initialization and
regularization (PaPIR), whereby the amount of knowledge transferred is
systemically adjusted for the initialization and regularization of the ANN
model in the target task. It is found that an additional performance gain can
be achieved by changing the initialization scheme of the ANN model in the
target task when the task similarity between source and target tasks is
relatively low.
Related papers
- Transfer Learning Based Hybrid Quantum Neural Network Model for Surface Anomaly Detection [0.4604003661048266]
This paper presents a quantum transfer learning (QTL) based approach to significantly reduce the number of parameters of the classical models.
We show that we could reduce the total number of trainable parameters up to 90% of the initial model without any drop in performance.
arXiv Detail & Related papers (2024-08-30T19:40:52Z) - chemtrain: Learning Deep Potential Models via Automatic Differentiation and Statistical Physics [0.0]
Neural Networks (NNs) are promising models for refining the accuracy of molecular dynamics.
Chemtrain is a framework to learn sophisticated NN potential models through customizable training routines and advanced training algorithms.
arXiv Detail & Related papers (2024-08-28T15:14:58Z) - Diffusion-Based Neural Network Weights Generation [80.89706112736353]
D2NWG is a diffusion-based neural network weights generation technique that efficiently produces high-performing weights for transfer learning.
Our method extends generative hyper-representation learning to recast the latent diffusion paradigm for neural network weights generation.
Our approach is scalable to large architectures such as large language models (LLMs), overcoming the limitations of current parameter generation techniques.
arXiv Detail & Related papers (2024-02-28T08:34:23Z) - Out of the Ordinary: Spectrally Adapting Regression for Covariate Shift [12.770658031721435]
We propose a method for adapting the weights of the last layer of a pre-trained neural regression model to perform better on input data originating from a different distribution.
We demonstrate how this lightweight spectral adaptation procedure can improve out-of-distribution performance for synthetic and real-world datasets.
arXiv Detail & Related papers (2023-12-29T04:15:58Z) - Diffusion-Model-Assisted Supervised Learning of Generative Models for
Density Estimation [10.793646707711442]
We present a framework for training generative models for density estimation.
We use the score-based diffusion model to generate labeled data.
Once the labeled data are generated, we can train a simple fully connected neural network to learn the generative model in the supervised manner.
arXiv Detail & Related papers (2023-10-22T23:56:19Z) - Diffusion Generative Flow Samplers: Improving learning signals through
partial trajectory optimization [87.21285093582446]
Diffusion Generative Flow Samplers (DGFS) is a sampling-based framework where the learning process can be tractably broken down into short partial trajectory segments.
Our method takes inspiration from the theory developed for generative flow networks (GFlowNets)
arXiv Detail & Related papers (2023-10-04T09:39:05Z) - AdaMerging: Adaptive Model Merging for Multi-Task Learning [68.75885518081357]
This paper introduces an innovative technique called Adaptive Model Merging (AdaMerging)
It aims to autonomously learn the coefficients for model merging, either in a task-wise or layer-wise manner, without relying on the original training data.
Compared to the current state-of-the-art task arithmetic merging scheme, AdaMerging showcases a remarkable 11% improvement in performance.
arXiv Detail & Related papers (2023-10-04T04:26:33Z) - Phasic Content Fusing Diffusion Model with Directional Distribution
Consistency for Few-Shot Model Adaption [73.98706049140098]
We propose a novel phasic content fusing few-shot diffusion model with directional distribution consistency loss.
Specifically, we design a phasic training strategy with phasic content fusion to help our model learn content and style information when t is large.
Finally, we propose a cross-domain structure guidance strategy that enhances structure consistency during domain adaptation.
arXiv Detail & Related papers (2023-09-07T14:14:11Z) - Quantized Adaptive Subgradient Algorithms and Their Applications [39.103587572626026]
We propose quantized composite mirror descent adaptive subgradient (QCMD adagrad) and quantized regularized dual average adaptive subgradient (QRDA adagrad) for distributed training.
A quantized gradient-based adaptive learning rate matrix is constructed to achieve a balance between communication costs, accuracy, and model sparsity.
arXiv Detail & Related papers (2022-08-11T04:04:03Z) - Churn Reduction via Distillation [54.5952282395487]
We show an equivalence between training with distillation using the base model as the teacher and training with an explicit constraint on the predictive churn.
We then show that distillation performs strongly for low churn training against a number of recent baselines.
arXiv Detail & Related papers (2021-06-04T18:03:31Z) - Towards Accurate Knowledge Transfer via Target-awareness Representation
Disentanglement [56.40587594647692]
We propose a novel transfer learning algorithm, introducing the idea of Target-awareness REpresentation Disentanglement (TRED)
TRED disentangles the relevant knowledge with respect to the target task from the original source model and used as a regularizer during fine-tuning the target model.
Experiments on various real world datasets show that our method stably improves the standard fine-tuning by more than 2% in average.
arXiv Detail & Related papers (2020-10-16T17:45:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.