High-Dimensional Yield Estimation using Shrinkage Deep Features and
Maximization of Integral Entropy Reduction
- URL: http://arxiv.org/abs/2212.02100v1
- Date: Mon, 5 Dec 2022 08:39:41 GMT
- Title: High-Dimensional Yield Estimation using Shrinkage Deep Features and
Maximization of Integral Entropy Reduction
- Authors: Shuo Yin, Guohao Dai, Wei W. Xing
- Abstract summary: We propose an absolute deep learning, ASDK, which automatically identifies the dominant process variation parameters in a nonlinear kernel-correlated deep kernel.
Experiments on column circuits demonstrate the superiority of ASDK over the state-of-the-art (SOTA) approaches in terms of accuracy and efficiency with up to 10.3x speedup over SOTA methods.
- Score: 0.8522010776600341
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite the fast advances in high-sigma yield analysis with the help of
machine learning techniques in the past decade, one of the main challenges, the
curse of dimensionality, which is inevitable when dealing with modern
large-scale circuits, remains unsolved. To resolve this challenge, we propose
an absolute shrinkage deep kernel learning, ASDK, which automatically
identifies the dominant process variation parameters in a nonlinear-correlated
deep kernel and acts as a surrogate model to emulate the expensive SPICE
simulation. To further improve the yield estimation efficiency, we propose a
novel maximization of approximated entropy reduction for an efficient model
update, which is also enhanced with parallel batch sampling for parallel
computing, making it ready for practical deployment. Experiments on SRAM column
circuits demonstrate the superiority of ASDK over the state-of-the-art (SOTA)
approaches in terms of accuracy and efficiency with up to 10.3x speedup over
SOTA methods.
Related papers
- Adaptive Anomaly Detection in Network Flows with Low-Rank Tensor Decompositions and Deep Unrolling [9.20186865054847]
Anomaly detection (AD) is increasingly recognized as a key component for ensuring the resilience of future communication systems.
This work considers AD in network flows using incomplete measurements.
We propose a novel block-successive convex approximation algorithm based on a regularized model-fitting objective.
Inspired by Bayesian approaches, we extend the model architecture to perform online adaptation to per-flow and per-time-step statistics.
arXiv Detail & Related papers (2024-09-17T19:59:57Z) - Efficient Learning With Sine-Activated Low-rank Matrices [25.12262017296922]
We propose a novel theoretical framework that integrates a sinusoidal function within the low-rank decomposition process.
Our method proves to be a plug in enhancement for existing low-rank models, as evidenced by its successful application in Vision Transformers (ViT), Large Language Models (LLMs), Neural Radiance Fields (NeRF) and 3D shape modelling.
arXiv Detail & Related papers (2024-03-28T08:58:20Z) - Seeking the Yield Barrier: High-Dimensional SRAM Evaluation Through
Optimal Manifold [3.258560324501261]
We develop a novel yield estimation method, named Optimal Manifold Important Sampling (OPTIMIS)
OPTIMIS delivers state-of-the-art performance with robustness and consistency, with up to 3.5x in efficiency and 3x in accuracy over the best of SOTA methods in High-dimensional evaluation.
arXiv Detail & Related papers (2023-07-28T19:21:39Z) - An Optimization-based Deep Equilibrium Model for Hyperspectral Image
Deconvolution with Convergence Guarantees [71.57324258813675]
We propose a novel methodology for addressing the hyperspectral image deconvolution problem.
A new optimization problem is formulated, leveraging a learnable regularizer in the form of a neural network.
The derived iterative solver is then expressed as a fixed-point calculation problem within the Deep Equilibrium framework.
arXiv Detail & Related papers (2023-06-10T08:25:16Z) - End-to-End Meta-Bayesian Optimisation with Transformer Neural Processes [52.818579746354665]
This paper proposes the first end-to-end differentiable meta-BO framework that generalises neural processes to learn acquisition functions via transformer architectures.
We enable this end-to-end framework with reinforcement learning (RL) to tackle the lack of labelled acquisition data.
arXiv Detail & Related papers (2023-05-25T10:58:46Z) - Conditional Denoising Diffusion for Sequential Recommendation [62.127862728308045]
Two prominent generative models, Generative Adversarial Networks (GANs) and Variational AutoEncoders (VAEs)
GANs suffer from unstable optimization, while VAEs are prone to posterior collapse and over-smoothed generations.
We present a conditional denoising diffusion model, which includes a sequence encoder, a cross-attentive denoising decoder, and a step-wise diffuser.
arXiv Detail & Related papers (2023-04-22T15:32:59Z) - Deep Equilibrium Optical Flow Estimation [80.80992684796566]
Recent state-of-the-art (SOTA) optical flow models use finite-step recurrent update operations to emulate traditional algorithms.
These RNNs impose large computation and memory overheads, and are not directly trained to model such stable estimation.
We propose deep equilibrium (DEQ) flow estimators, an approach that directly solves for the flow as the infinite-level fixed point of an implicit layer.
arXiv Detail & Related papers (2022-04-18T17:53:44Z) - Generalised Latent Assimilation in Heterogeneous Reduced Spaces with
Machine Learning Surrogate Models [10.410970649045943]
We develop a system which combines reduced-order surrogate models with a novel data assimilation technique.
Generalised Latent Assimilation can benefit both the efficiency provided by the reduced-order modelling and the accuracy of data assimilation.
arXiv Detail & Related papers (2022-04-07T15:13:12Z) - Fast Distributionally Robust Learning with Variance Reduced Min-Max
Optimization [85.84019017587477]
Distributionally robust supervised learning is emerging as a key paradigm for building reliable machine learning systems for real-world applications.
Existing algorithms for solving Wasserstein DRSL involve solving complex subproblems or fail to make use of gradients.
We revisit Wasserstein DRSL through the lens of min-max optimization and derive scalable and efficiently implementable extra-gradient algorithms.
arXiv Detail & Related papers (2021-04-27T16:56:09Z) - Extrapolation for Large-batch Training in Deep Learning [72.61259487233214]
We show that a host of variations can be covered in a unified framework that we propose.
We prove the convergence of this novel scheme and rigorously evaluate its empirical performance on ResNet, LSTM, and Transformer.
arXiv Detail & Related papers (2020-06-10T08:22:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.