Accelerated and Inexpensive Machine Learning for Manufacturing Processes
with Incomplete Mechanistic Knowledge
- URL: http://arxiv.org/abs/2305.00229v1
- Date: Sat, 29 Apr 2023 10:54:57 GMT
- Title: Accelerated and Inexpensive Machine Learning for Manufacturing Processes
with Incomplete Mechanistic Knowledge
- Authors: Jeremy Cleeman, Kian Agrawala, Rajiv Malhotra
- Abstract summary: This paper proposes a transfer learning based approach to address this issue.
A ML model is trained on a large amount of computationally inexpensive data from a physics-based process model (source) and then fine-tuned on a smaller amount of costly experimental data (target)
Despite extreme functional and quantitative inaccuracies in the source our approach reduces the model development cost by years, experimental cost by 56-76%, computational cost by orders of magnitude, and prediction error by 16-24%.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Machine Learning (ML) is of increasing interest for modeling parametric
effects in manufacturing processes. But this approach is limited to established
processes for which a deep physics-based understanding has been developed over
time, since state-of-the-art approaches focus on reducing the experimental
and/or computational costs of generating the training data but ignore the
inherent and significant cost of developing qualitatively accurate
physics-based models for new processes . This paper proposes a transfer
learning based approach to address this issue, in which a ML model is trained
on a large amount of computationally inexpensive data from a physics-based
process model (source) and then fine-tuned on a smaller amount of costly
experimental data (target). The novelty lies in pushing the boundaries of the
qualitative accuracy demanded of the source model, which is assumed to be high
in the literature, and is the root of the high model development cost. Our
approach is evaluated for modeling the printed line width in Fused Filament
Fabrication. Despite extreme functional and quantitative inaccuracies in the
source our approach reduces the model development cost by years, experimental
cost by 56-76%, computational cost by orders of magnitude, and prediction error
by 16-24%.
Related papers
- Rare Class Prediction Model for Smart Industry in Semiconductor Manufacturing [1.3955252961896323]
This study develops a rare class prediction approach for in situ data collected from a smart semiconductor manufacturing process.
The primary objective is to build a model that addresses issues of noise and class imbalance, enhancing class separation.
The model was evaluated using various performance metrics, with ROC curves showing an AUC of 0.95, a precision of 0.66, and a recall of 0.96
arXiv Detail & Related papers (2024-06-06T22:09:43Z) - Multifidelity linear regression for scientific machine learning from scarce data [0.0]
We propose a new multifidelity training approach for scientific machine learning via linear regression.
We provide bias and variance analysis of our new estimators that guarantee the approach's accuracy and improved robustness to scarce high-fidelity data.
arXiv Detail & Related papers (2024-03-13T15:40:17Z) - QualEval: Qualitative Evaluation for Model Improvement [82.73561470966658]
We propose QualEval, which augments quantitative scalar metrics with automated qualitative evaluation as a vehicle for model improvement.
QualEval uses a powerful LLM reasoner and our novel flexible linear programming solver to generate human-readable insights.
We demonstrate that leveraging its insights, for example, improves the absolute performance of the Llama 2 model by up to 15% points relative.
arXiv Detail & Related papers (2023-11-06T00:21:44Z) - Reusing Pretrained Models by Multi-linear Operators for Efficient
Training [65.64075958382034]
Training large models from scratch usually costs a substantial amount of resources.
Recent studies such as bert2BERT and LiGO have reused small pretrained models to initialize a large model.
We propose a method that linearly correlates each weight of the target model to all the weights of the pretrained model.
arXiv Detail & Related papers (2023-10-16T06:16:47Z) - Cost-Effective Retraining of Machine Learning Models [2.9461360639852914]
It is important to retrain a machine learning (ML) model in order to maintain its performance as the data changes over time.
This creates a trade-off between retraining too frequently, which leads to unnecessary computing costs, and not retraining often enough.
We propose ML systems that make automated and cost-effective decisions about when to retrain an ML model.
arXiv Detail & Related papers (2023-10-06T13:02:29Z) - Uncertainty Estimation in Machine Learning [0.0]
In machine learning the model complexity and severe nonlinearity become serious obstacles to uncertainty evaluation.
The latest example of a pre-trained model is the Generative Pre-trained Transformer 3 with hundreds of billions of parameters and a half-terabyte training dataset.
arXiv Detail & Related papers (2022-06-03T16:11:11Z) - Alternate Model Growth and Pruning for Efficient Training of
Recommendation Systems [7.415129876303651]
Model pruning is an effective technique to reduce computation overhead for deep neural networks by removing redundant parameters.
Modern recommendation systems are still thirsty for model capacity due to the demand for handling big data.
We propose a dynamic training scheme, namely alternate model growth and pruning, to alternatively construct and prune weights in the course of training.
arXiv Detail & Related papers (2021-05-04T03:14:30Z) - Model-Augmented Q-learning [112.86795579978802]
We propose a MFRL framework that is augmented with the components of model-based RL.
Specifically, we propose to estimate not only the $Q$-values but also both the transition and the reward with a shared network.
We show that the proposed scheme, called Model-augmented $Q$-learning (MQL), obtains a policy-invariant solution which is identical to the solution obtained by learning with true reward.
arXiv Detail & Related papers (2021-02-07T17:56:50Z) - Transfer Learning without Knowing: Reprogramming Black-box Machine
Learning Models with Scarce Data and Limited Resources [78.72922528736011]
We propose a novel approach, black-box adversarial reprogramming (BAR), that repurposes a well-trained black-box machine learning model.
Using zeroth order optimization and multi-label mapping techniques, BAR can reprogram a black-box ML model solely based on its input-output responses.
BAR outperforms state-of-the-art methods and yields comparable performance to the vanilla adversarial reprogramming method.
arXiv Detail & Related papers (2020-07-17T01:52:34Z) - VAE-LIME: Deep Generative Model Based Approach for Local Data-Driven
Model Interpretability Applied to the Ironmaking Industry [70.10343492784465]
It is necessary to expose to the process engineer, not solely the model predictions, but also their interpretability.
Model-agnostic local interpretability solutions based on LIME have recently emerged to improve the original method.
We present in this paper a novel approach, VAE-LIME, for local interpretability of data-driven models forecasting the temperature of the hot metal produced by a blast furnace.
arXiv Detail & Related papers (2020-07-15T07:07:07Z) - Predictive modeling approaches in laser-based material processing [59.04160452043105]
This study aims to automate and forecast the effect of laser processing on material structures.
The focus is centred on the performance of representative statistical and machine learning algorithms.
Results can set the basis for a systematic methodology towards reducing material design, testing and production cost.
arXiv Detail & Related papers (2020-06-13T17:28:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.