Deep Learning to Automate Parameter Extraction and Model Fitting of Two-Dimensional Transistors
- URL: http://arxiv.org/abs/2507.05134v1
- Date: Mon, 07 Jul 2025 15:46:25 GMT
- Title: Deep Learning to Automate Parameter Extraction and Model Fitting of Two-Dimensional Transistors
- Authors: Robert K. A. Bennett, Jan-Lucas Uslu, Harmon F. Gault, Asir Intisar Khan, Lauren Hoang, Tara Peña, Kathryn Neilson, Young Suh Song, Zhepeng Zhang, Andrew J. Mannix, Eric Pop,
- Abstract summary: We present a deep learning approach to extract physical parameters of 2D transistors from electrical measurements.<n>We train a secondary neural network to approximate a physics-based device simulator.<n>This method enables high-quality fits after training the neural network on electrical data generated from simulations of 500 devices.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present a deep learning approach to extract physical parameters (e.g., mobility, Schottky contact barrier height, defect profiles) of two-dimensional (2D) transistors from electrical measurements, enabling automated parameter extraction and technology computer-aided design (TCAD) fitting. To facilitate this task, we implement a simple data augmentation and pre-training approach by training a secondary neural network to approximate a physics-based device simulator. This method enables high-quality fits after training the neural network on electrical data generated from physics-based simulations of ~500 devices, a factor >40$\times$ fewer than other recent efforts. Consequently, fitting can be achieved by training on physically rigorous TCAD models, including complex geometry, self-consistent transport, and electrostatic effects, and is not limited to computationally inexpensive compact models. We apply our approach to reverse-engineer key parameters from experimental monolayer WS$_2$ transistors, achieving a median coefficient of determination ($R^2$) = 0.99 when fitting measured electrical data. We also demonstrate that this approach generalizes and scales well by reverse-engineering electrical data on high-electron-mobility transistors while fitting 35 parameters simultaneously. To facilitate future research on deep learning approaches for inverse transistor design, we have published our code and sample data sets online.
Related papers
- Fusing CFD and measurement data using transfer learning [49.1574468325115]
We introduce a non-linear method based on neural networks combining simulation and measurement data via transfer learning.<n>In a first step, the neural network is trained on simulation data to learn spatial features of the distributed quantities.<n>The second step involves transfer learning on the measurement data to correct for systematic errors between simulation and measurement by only re-training a small subset of the entire neural network model.
arXiv Detail & Related papers (2025-07-28T07:21:46Z) - Ga$_2$O$_3$ TCAD Mobility Parameter Calibration using Simulation Augmented Machine Learning with Physics Informed Neural Network [3.194221922047046]
We show the possibility of performing automatic Technology Computer-Aided-Design (TCAD) parameter calibration using machine learning, verified with experimental data.<n>A machine comprised of an autoencoder (AE) and a neural network (NN) (AE-NN) is used.<n>TCAD extracted parameters shows that the quality of the parameters is as good as an expert at the pre-turned-on regime but not in the on-state regime.
arXiv Detail & Related papers (2025-04-03T05:09:43Z) - Physically Guided Deep Unsupervised Inversion for 1D Magnetotelluric Models [16.91835461818938]
This paper presents a new deep unsupervised inversion algorithm guided by physics to estimate 1D Magnetotelluric (MT) models.<n>Instead of using datasets with the observed data and their respective models as labels during training, our method employs a differentiable modeling operator that physically guides the cost function minimization.
arXiv Detail & Related papers (2024-10-20T04:17:59Z) - Transformer neural networks and quantum simulators: a hybrid approach for simulating strongly correlated systems [1.6494451064539348]
We present a hybrid optimization scheme for neural quantum states (NQS)<n>By using both projective measurements from the computational basis as well as expectation values from other measurement configurations, our pretraining gives access to the sign structure of the state.<n>Our work paves the way for a reliable and efficient optimization of neural quantum states.
arXiv Detail & Related papers (2024-05-31T17:55:27Z) - Learning-based adaption of robotic friction models [50.72489248401199]
We introduce a novel approach to adapt an existing friction model to new dynamics using as little data as possible.<n>Our method does not rely on data with external load during training, eliminating the need for external torque sensors.
arXiv Detail & Related papers (2023-10-25T14:50:15Z) - Improved particle-flow event reconstruction with scalable neural networks for current and future particle detectors [1.4609888393206634]
We study scalable machine learning models for event reconstruction in electron-positron collisions based on a full detector simulation.
We compare a graph neural network and kernel-based transformer and demonstrate that we can avoid operations while achieving realistic reconstruction.
The best graph neural network model shows improvement in the jet transverse momentum resolution by up to 50% compared to the rule-based algorithm.
arXiv Detail & Related papers (2023-09-13T08:16:15Z) - Efficient GPT Model Pre-training using Tensor Train Matrix
Representation [65.96485282393361]
Large-scale transformer models feature billions of parameters, leading to difficulties in their deployment and prohibitive training costs from scratch.
To reduce the number of parameters in the GPT-2 architecture, we replace the matrices of fully-connected layers with the corresponding Train Matrix(TTM) structure.
The resulting GPT-based model stores up to 40% fewer parameters, showing the perplexity comparable to the original model.
arXiv Detail & Related papers (2023-06-05T08:38:25Z) - Pretraining Graph Neural Networks for few-shot Analog Circuit Modeling
and Design [68.1682448368636]
We present a supervised pretraining approach to learn circuit representations that can be adapted to new unseen topologies or unseen prediction tasks.
To cope with the variable topological structure of different circuits we describe each circuit as a graph and use graph neural networks (GNNs) to learn node embeddings.
We show that pretraining GNNs on prediction of output node voltages can encourage learning representations that can be adapted to new unseen topologies or prediction of new circuit level properties.
arXiv Detail & Related papers (2022-03-29T21:18:47Z) - Semi-supervised physics guided deep learning framework for predicting
the I-V characteristics of GAN HEMT [0.0]
The framework is generic in nature and can be applied to model a phenomenon from other fields of research too as long as its behaviour is known.
A semi-supervised physics guided neural network (SPGNN) has been developed that predicts I-V characteristics of a gallium nitride-based high electron mobility transistor (GaN HEMT)
The SPGNN significantly reduces the requirement of the training data by more than 80% for achieving similar or better performance than a traditional neural network (TNN) even for unseen conditions.
arXiv Detail & Related papers (2021-10-20T18:48:50Z) - Efficient pre-training objectives for Transformers [84.64393460397471]
We study several efficient pre-training objectives for Transformers-based models.
We prove that eliminating the MASK token and considering the whole output during the loss are essential choices to improve performance.
arXiv Detail & Related papers (2021-04-20T00:09:37Z) - Thermal Neural Networks: Lumped-Parameter Thermal Modeling With
State-Space Machine Learning [0.0]
Thermal models for electric power systems are required to be both, real-time capable and of high estimation accuracy.
In this work, the thermal neural network (TNN) is introduced, which unifies both, consolidated knowledge in the form of heat-transfer-based lumped- parameter models.
A TNN has physically interpretable states through its state-space representation, is end-to-end trainable, and requires no material, geometry, nor expert knowledge for its design.
arXiv Detail & Related papers (2021-03-30T13:15:48Z) - Machine learning for rapid discovery of laminar flow channel wall
modifications that enhance heat transfer [56.34005280792013]
We present a combination of accurate numerical simulations of arbitrary, flat, and non-flat channels and machine learning models predicting drag coefficient and Stanton number.
We show that convolutional neural networks (CNN) can accurately predict the target properties at a fraction of the time of numerical simulations.
arXiv Detail & Related papers (2021-01-19T16:14:02Z) - One-step regression and classification with crosspoint resistive memory
arrays [62.997667081978825]
High speed, low energy computing machines are in demand to enable real-time artificial intelligence at the edge.
One-step learning is supported by simulations of the prediction of the cost of a house in Boston and the training of a 2-layer neural network for MNIST digit recognition.
Results are all obtained in one computational step, thanks to the physical, parallel, and analog computing within the crosspoint array.
arXiv Detail & Related papers (2020-05-05T08:00:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.