Optimizing Predictive AI in Physical Design Flows with Mini Pixel Batch
Gradient Descent
- URL: http://arxiv.org/abs/2402.06034v1
- Date: Thu, 8 Feb 2024 20:14:35 GMT
- Title: Optimizing Predictive AI in Physical Design Flows with Mini Pixel Batch
Gradient Descent
- Authors: Haoyu Yang and Anthony Agnesina and Haoxing Ren
- Abstract summary: We argue the averaging effect of MSE induces limitations in both model training and deployment.
We propose mini-pixel gradient batch descent (MPGD), a plug-and-play optimization algorithm.
Experiments on representative benchmark suits show the significant benefits of MPGD on various physical design prediction tasks.
- Score: 5.413212114044892
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Exploding predictive AI has enabled fast yet effective evaluation and
decision-making in modern chip physical design flows. State-of-the-art
frameworks typically include the objective of minimizing the mean square error
(MSE) between the prediction and the ground truth. We argue the averaging
effect of MSE induces limitations in both model training and deployment, and
good MSE behavior does not guarantee the capability of these models to assist
physical design flows which are likely sabotaged due to a small portion of
prediction error. To address this, we propose mini-pixel batch gradient descent
(MPGD), a plug-and-play optimization algorithm that takes the most informative
entries into consideration, offering probably faster and better convergence.
Experiments on representative benchmark suits show the significant benefits of
MPGD on various physical design prediction tasks using CNN or Graph-based
models.
Related papers
- Cliqueformer: Model-Based Optimization with Structured Transformers [102.55764949282906]
We develop a model that learns the structure of an MBO task and empirically leads to improved designs.
We evaluate Cliqueformer on various tasks, ranging from high-dimensional black-box functions to real-world tasks of chemical and genetic design.
arXiv Detail & Related papers (2024-10-17T00:35:47Z) - Revisiting SMoE Language Models by Evaluating Inefficiencies with Task Specific Expert Pruning [78.72226641279863]
Sparse Mixture of Expert (SMoE) models have emerged as a scalable alternative to dense models in language modeling.
Our research explores task-specific model pruning to inform decisions about designing SMoE architectures.
We introduce an adaptive task-aware pruning technique UNCURL to reduce the number of experts per MoE layer in an offline manner post-training.
arXiv Detail & Related papers (2024-09-02T22:35:03Z) - Optimization of DNN-based speaker verification model through efficient quantization technique [15.250677730668466]
Quantization of deep models offers a means to reduce both computational and memory expenses.
Our research proposes an optimization framework for the quantization of the speaker verification model.
arXiv Detail & Related papers (2024-07-12T05:03:10Z) - On Robustness and Generalization of ML-Based Congestion Predictors to
Valid and Imperceptible Perturbations [9.982978359852494]
Recent work has demonstrated that neural networks are generally vulnerable to small, carefully chosen perturbations of their input.
We show that state-of-the-art CNN and GNN-based congestion models exhibit brittleness to imperceptible perturbations.
Our work indicates that CAD engineers should be cautious when integrating neural network-based mechanisms in EDA flows.
arXiv Detail & Related papers (2024-02-29T20:11:47Z) - Evaluation of Differentially Constrained Motion Models for Graph-Based
Trajectory Prediction [1.1947990549568765]
This research investigates the performance of various motion models in combination with numerical solvers for the prediction task.
The study shows that simpler models, such as low-order integrator models, are preferred over more complex, e.g., kinematic models, to achieve accurate predictions.
arXiv Detail & Related papers (2023-04-11T10:15:20Z) - When to Update Your Model: Constrained Model-based Reinforcement
Learning [50.74369835934703]
We propose a novel and general theoretical scheme for a non-decreasing performance guarantee of model-based RL (MBRL)
Our follow-up derived bounds reveal the relationship between model shifts and performance improvement.
A further example demonstrates that learning models from a dynamically-varying number of explorations benefit the eventual returns.
arXiv Detail & Related papers (2022-10-15T17:57:43Z) - AMED: Automatic Mixed-Precision Quantization for Edge Devices [3.5223695602582614]
Quantized neural networks are well known for reducing the latency, power consumption, and model size without significant harm to the performance.
Mixed-precision quantization offers better utilization of customized hardware that supports arithmetic operations at different bitwidths.
arXiv Detail & Related papers (2022-05-30T21:23:22Z) - Interpretable AI-based Large-scale 3D Pathloss Prediction Model for
enabling Emerging Self-Driving Networks [3.710841042000923]
We propose a Machine Learning-based model that leverages novel key predictors for estimating pathloss.
By quantitatively evaluating the ability of various ML algorithms in terms of predictive, generalization and computational performance, our results show that Light Gradient Boosting Machine (LightGBM) algorithm overall outperforms others.
arXiv Detail & Related papers (2022-01-30T19:50:16Z) - Conservative Objective Models for Effective Offline Model-Based
Optimization [78.19085445065845]
Computational design problems arise in a number of settings, from synthetic biology to computer architectures.
We propose a method that learns a model of the objective function that lower bounds the actual value of the ground-truth objective on out-of-distribution inputs.
COMs are simple to implement and outperform a number of existing methods on a wide range of MBO problems.
arXiv Detail & Related papers (2021-07-14T17:55:28Z) - Covert Model Poisoning Against Federated Learning: Algorithm Design and
Optimization [76.51980153902774]
Federated learning (FL) is vulnerable to external attacks on FL models during parameters transmissions.
In this paper, we propose effective MP algorithms to combat state-of-the-art defensive aggregation mechanisms.
Our experimental results demonstrate that the proposed CMP algorithms are effective and substantially outperform existing attack mechanisms.
arXiv Detail & Related papers (2021-01-28T03:28:18Z) - Models, Pixels, and Rewards: Evaluating Design Trade-offs in Visual
Model-Based Reinforcement Learning [109.74041512359476]
We study a number of design decisions for the predictive model in visual MBRL algorithms.
We find that a range of design decisions that are often considered crucial, such as the use of latent spaces, have little effect on task performance.
We show how this phenomenon is related to exploration and how some of the lower-scoring models on standard benchmarks will perform the same as the best-performing models when trained on the same training data.
arXiv Detail & Related papers (2020-12-08T18:03:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.