Accelerated and Inexpensive Machine Learning for Manufacturing Processes
with Incomplete Mechanistic Knowledge
- URL: http://arxiv.org/abs/2305.00229v1
- Date: Sat, 29 Apr 2023 10:54:57 GMT
- Title: Accelerated and Inexpensive Machine Learning for Manufacturing Processes
with Incomplete Mechanistic Knowledge
- Authors: Jeremy Cleeman, Kian Agrawala, Rajiv Malhotra
- Abstract summary: This paper proposes a transfer learning based approach to address this issue.
A ML model is trained on a large amount of computationally inexpensive data from a physics-based process model (source) and then fine-tuned on a smaller amount of costly experimental data (target)
Despite extreme functional and quantitative inaccuracies in the source our approach reduces the model development cost by years, experimental cost by 56-76%, computational cost by orders of magnitude, and prediction error by 16-24%.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Machine Learning (ML) is of increasing interest for modeling parametric
effects in manufacturing processes. But this approach is limited to established
processes for which a deep physics-based understanding has been developed over
time, since state-of-the-art approaches focus on reducing the experimental
and/or computational costs of generating the training data but ignore the
inherent and significant cost of developing qualitatively accurate
physics-based models for new processes . This paper proposes a transfer
learning based approach to address this issue, in which a ML model is trained
on a large amount of computationally inexpensive data from a physics-based
process model (source) and then fine-tuned on a smaller amount of costly
experimental data (target). The novelty lies in pushing the boundaries of the
qualitative accuracy demanded of the source model, which is assumed to be high
in the literature, and is the root of the high model development cost. Our
approach is evaluated for modeling the printed line width in Fused Filament
Fabrication. Despite extreme functional and quantitative inaccuracies in the
source our approach reduces the model development cost by years, experimental
cost by 56-76%, computational cost by orders of magnitude, and prediction error
by 16-24%.
Related papers
- Computation-Aware Gaussian Processes: Model Selection And Linear-Time Inference [55.150117654242706]
We show that model selection for computation-aware GPs trained on 1.8 million data points can be done within a few hours on a single GPU.
As a result of this work, Gaussian processes can be trained on large-scale datasets without significantly compromising their ability to quantify uncertainty.
arXiv Detail & Related papers (2024-11-01T21:11:48Z) - Transfer Learning on Multi-Dimensional Data: A Novel Approach to Neural Network-Based Surrogate Modeling [0.0]
Convolutional neural networks (CNNs) have gained popularity as the basis for such surrogate models.
We propose training a CNN surrogate model on a mixture of numerical solutions to both the $d$-dimensional problem and its ($d-1$)-dimensional approximation.
We demonstrate our approach on a multiphase flow test problem, using transfer learning to train a dense fully-convolutional encoder-decoder CNN on the two classes of data.
arXiv Detail & Related papers (2024-10-16T05:07:48Z) - Investigating Data Hierarchies in Multifidelity Machine Learning for Excitation Energies [0.0]
This study investigates the impact of modifying $gamma$ on model efficiency and accuracy for the prediction of vertical excitation energies using the QeMFi benchmark dataset.
A novel error metric, error contours of MFML, is proposed to provide a comprehensive view of model error contributions from each fidelity.
The results indicate that high model accuracy can be achieved with just 2 training samples at the target fidelity when a larger number of samples from lower fidelities are used.
arXiv Detail & Related papers (2024-10-15T08:35:00Z) - Multifidelity linear regression for scientific machine learning from scarce data [0.0]
We propose a new multifidelity training approach for scientific machine learning via linear regression.
We provide bias and variance analysis of our new estimators that guarantee the approach's accuracy and improved robustness to scarce high-fidelity data.
arXiv Detail & Related papers (2024-03-13T15:40:17Z) - QualEval: Qualitative Evaluation for Model Improvement [82.73561470966658]
We propose QualEval, which augments quantitative scalar metrics with automated qualitative evaluation as a vehicle for model improvement.
QualEval uses a powerful LLM reasoner and our novel flexible linear programming solver to generate human-readable insights.
We demonstrate that leveraging its insights, for example, improves the absolute performance of the Llama 2 model by up to 15% points relative.
arXiv Detail & Related papers (2023-11-06T00:21:44Z) - Uncertainty Estimation in Machine Learning [0.0]
In machine learning the model complexity and severe nonlinearity become serious obstacles to uncertainty evaluation.
The latest example of a pre-trained model is the Generative Pre-trained Transformer 3 with hundreds of billions of parameters and a half-terabyte training dataset.
arXiv Detail & Related papers (2022-06-03T16:11:11Z) - Alternate Model Growth and Pruning for Efficient Training of
Recommendation Systems [7.415129876303651]
Model pruning is an effective technique to reduce computation overhead for deep neural networks by removing redundant parameters.
Modern recommendation systems are still thirsty for model capacity due to the demand for handling big data.
We propose a dynamic training scheme, namely alternate model growth and pruning, to alternatively construct and prune weights in the course of training.
arXiv Detail & Related papers (2021-05-04T03:14:30Z) - Model-Augmented Q-learning [112.86795579978802]
We propose a MFRL framework that is augmented with the components of model-based RL.
Specifically, we propose to estimate not only the $Q$-values but also both the transition and the reward with a shared network.
We show that the proposed scheme, called Model-augmented $Q$-learning (MQL), obtains a policy-invariant solution which is identical to the solution obtained by learning with true reward.
arXiv Detail & Related papers (2021-02-07T17:56:50Z) - Transfer Learning without Knowing: Reprogramming Black-box Machine
Learning Models with Scarce Data and Limited Resources [78.72922528736011]
We propose a novel approach, black-box adversarial reprogramming (BAR), that repurposes a well-trained black-box machine learning model.
Using zeroth order optimization and multi-label mapping techniques, BAR can reprogram a black-box ML model solely based on its input-output responses.
BAR outperforms state-of-the-art methods and yields comparable performance to the vanilla adversarial reprogramming method.
arXiv Detail & Related papers (2020-07-17T01:52:34Z) - VAE-LIME: Deep Generative Model Based Approach for Local Data-Driven
Model Interpretability Applied to the Ironmaking Industry [70.10343492784465]
It is necessary to expose to the process engineer, not solely the model predictions, but also their interpretability.
Model-agnostic local interpretability solutions based on LIME have recently emerged to improve the original method.
We present in this paper a novel approach, VAE-LIME, for local interpretability of data-driven models forecasting the temperature of the hot metal produced by a blast furnace.
arXiv Detail & Related papers (2020-07-15T07:07:07Z) - Predictive modeling approaches in laser-based material processing [59.04160452043105]
This study aims to automate and forecast the effect of laser processing on material structures.
The focus is centred on the performance of representative statistical and machine learning algorithms.
Results can set the basis for a systematic methodology towards reducing material design, testing and production cost.
arXiv Detail & Related papers (2020-06-13T17:28:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.