Soft Sensing Model Visualization: Fine-tuning Neural Network from What
Model Learned
- URL: http://arxiv.org/abs/2111.06982v1
- Date: Fri, 12 Nov 2021 23:32:06 GMT
- Title: Soft Sensing Model Visualization: Fine-tuning Neural Network from What
Model Learned
- Authors: Xiaoye Qian, Chao Zhang, Jaswanth Yella, Yu Huang, Ming-Chun Huang,
Sthitie Bom
- Abstract summary: Data-driven soft-sensing modeling has become more prevalent in wafer process diagnostics.
Deep learning has been utilized in soft sensing system with promising performance on highly nonlinear and dynamic time-series data.
In this paper, we propose a deep learning-based model for defective wafer detection using a highly imbalanced dataset.
- Score: 5.182947614447375
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The growing availability of the data collected from smart manufacturing is
changing the paradigms of production monitoring and control. The increasing
complexity and content of the wafer manufacturing process in addition to the
time-varying unexpected disturbances and uncertainties, make it infeasible to
do the control process with model-based approaches. As a result, data-driven
soft-sensing modeling has become more prevalent in wafer process diagnostics.
Recently, deep learning has been utilized in soft sensing system with promising
performance on highly nonlinear and dynamic time-series data. Despite its
successes in soft-sensing systems, however, the underlying logic of the deep
learning framework is hard to understand. In this paper, we propose a deep
learning-based model for defective wafer detection using a highly imbalanced
dataset. To understand how the proposed model works, the deep visualization
approach is applied. Additionally, the model is then fine-tuned guided by the
deep visualization. Extensive experiments are performed to validate the
effectiveness of the proposed system. The results provide an interpretation of
how the model works and an instructive fine-tuning method based on the
interpretation.
Related papers
- Unlearning Information Bottleneck: Machine Unlearning of Systematic Patterns and Biases [6.936871609178494]
We present Unlearning Information Bottleneck (UIB), a novel information-theoretic framework designed to enhance the process of machine unlearning.
By proposing a variational upper bound, we recalibrate the model parameters through a dynamic prior that integrates changes in data distribution with an affordable computational cost.
Our experiments across various datasets, models, and unlearning methods demonstrate that our approach effectively removes systematic patterns and biases while maintaining the performance of models post-unlearning.
arXiv Detail & Related papers (2024-05-22T21:54:05Z) - DetDiffusion: Synergizing Generative and Perceptive Models for Enhanced Data Generation and Perception [78.26734070960886]
Current perceptive models heavily depend on resource-intensive datasets.
We introduce perception-aware loss (P.A. loss) through segmentation, improving both quality and controllability.
Our method customizes data augmentation by extracting and utilizing perception-aware attribute (P.A. Attr) during generation.
arXiv Detail & Related papers (2024-03-20T04:58:03Z) - Enhancing Dynamical System Modeling through Interpretable Machine
Learning Augmentations: A Case Study in Cathodic Electrophoretic Deposition [0.8796261172196743]
We introduce a comprehensive data-driven framework aimed at enhancing the modeling of physical systems.
As a demonstrative application, we pursue the modeling of cathodic electrophoretic deposition (EPD), commonly known as e-coating.
arXiv Detail & Related papers (2024-01-16T14:58:21Z) - Neural Harmonium: An Interpretable Deep Structure for Nonlinear Dynamic
System Identification with Application to Audio Processing [4.599180419117645]
Interpretability helps us understand a model's ability to generalize and reveal its limitations.
We introduce a causal interpretable deep structure for modeling dynamic systems.
Our proposed model makes use of the harmonic analysis by modeling the system in a time-frequency domain.
arXiv Detail & Related papers (2023-10-10T21:32:15Z) - Loop Polarity Analysis to Avoid Underspecification in Deep Learning [0.0]
In this paper, we turn to loop polarity analysis as a tool for specifying the causal structure of a data-generating process.
We show how measuring the polarity of the different feedback loops that compose a system can lead to more robust inferences on the part of neural networks.
arXiv Detail & Related papers (2023-09-18T23:49:42Z) - Robustness and Generalization Performance of Deep Learning Models on
Cyber-Physical Systems: A Comparative Study [71.84852429039881]
Investigation focuses on the models' ability to handle a range of perturbations, such as sensor faults and noise.
We test the generalization and transfer learning capabilities of these models by exposing them to out-of-distribution (OOD) samples.
arXiv Detail & Related papers (2023-06-13T12:43:59Z) - BOOT: Data-free Distillation of Denoising Diffusion Models with
Bootstrapping [64.54271680071373]
Diffusion models have demonstrated excellent potential for generating diverse images.
Knowledge distillation has been recently proposed as a remedy that can reduce the number of inference steps to one or a few.
We present a novel technique called BOOT, that overcomes limitations with an efficient data-free distillation algorithm.
arXiv Detail & Related papers (2023-06-08T20:30:55Z) - Ultrasound Signal Processing: From Models to Deep Learning [64.56774869055826]
Medical ultrasound imaging relies heavily on high-quality signal processing to provide reliable and interpretable image reconstructions.
Deep learning based methods, which are optimized in a data-driven fashion, have gained popularity.
A relatively new paradigm combines the power of the two: leveraging data-driven deep learning, as well as exploiting domain knowledge.
arXiv Detail & Related papers (2022-04-09T13:04:36Z) - Leveraging the structure of dynamical systems for data-driven modeling [111.45324708884813]
We consider the impact of the training set and its structure on the quality of the long-term prediction.
We show how an informed design of the training set, based on invariants of the system and the structure of the underlying attractor, significantly improves the resulting models.
arXiv Detail & Related papers (2021-12-15T20:09:20Z) - Bridging the Gap: Machine Learning to Resolve Improperly Modeled
Dynamics [4.940323406667406]
We present a data-driven modeling strategy to overcome improperly modeled dynamics for systems exhibiting complex-temporal behaviors.
We propose a Deep Learning framework to resolve the differences between the true dynamics of the system and the dynamics given by a model of the system that is either inaccurately or inadequately described.
arXiv Detail & Related papers (2020-08-23T04:57:02Z) - How Training Data Impacts Performance in Learning-based Control [67.7875109298865]
This paper derives an analytical relationship between the density of the training data and the control performance.
We formulate a quality measure for the data set, which we refer to as $rho$-gap.
We show how the $rho$-gap can be applied to a feedback linearizing control law.
arXiv Detail & Related papers (2020-05-25T12:13:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.