Pretraining Graph Neural Networks for few-shot Analog Circuit Modeling
and Design
- URL: http://arxiv.org/abs/2203.15913v1
- Date: Tue, 29 Mar 2022 21:18:47 GMT
- Title: Pretraining Graph Neural Networks for few-shot Analog Circuit Modeling
and Design
- Authors: Kourosh Hakhamaneshi, Marcel Nassar, Mariano Phielipp, Pieter Abbeel,
Vladimir Stojanovi\'c
- Abstract summary: We present a supervised pretraining approach to learn circuit representations that can be adapted to new unseen topologies or unseen prediction tasks.
To cope with the variable topological structure of different circuits we describe each circuit as a graph and use graph neural networks (GNNs) to learn node embeddings.
We show that pretraining GNNs on prediction of output node voltages can encourage learning representations that can be adapted to new unseen topologies or prediction of new circuit level properties.
- Score: 68.1682448368636
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Being able to predict the performance of circuits without running expensive
simulations is a desired capability that can catalyze automated design. In this
paper, we present a supervised pretraining approach to learn circuit
representations that can be adapted to new circuit topologies or unseen
prediction tasks. We hypothesize that if we train a neural network (NN) that
can predict the output DC voltages of a wide range of circuit instances it will
be forced to learn generalizable knowledge about the role of each circuit
element and how they interact with each other. The dataset for this supervised
learning objective can be easily collected at scale since the required DC
simulation to get ground truth labels is relatively cheap. This representation
would then be helpful for few-shot generalization to unseen circuit metrics
that require more time consuming simulations for obtaining the ground-truth
labels. To cope with the variable topological structure of different circuits
we describe each circuit as a graph and use graph neural networks (GNNs) to
learn node embeddings. We show that pretraining GNNs on prediction of output
node voltages can encourage learning representations that can be adapted to new
unseen topologies or prediction of new circuit level properties with up to 10x
more sample efficiency compared to a randomly initialized model. We further
show that we can improve sample efficiency of prior SoTA model-based
optimization methods by 2x (almost as good as using an oracle model) via
fintuning pretrained GNNs as the feature extractor of the learned models.
Related papers
- Learning from Linear Algebra: A Graph Neural Network Approach to Preconditioner Design for Conjugate Gradient Solvers [42.69799418639716]
Deep learning models may be used to precondition residuals during iteration of such linear solvers as the conjugate gradient (CG) method.
Neural network models require an enormous number of parameters to approximate well in this setup.
In our work, we recall well-established preconditioners from linear algebra and use them as a starting point for training the GNN.
arXiv Detail & Related papers (2024-05-24T13:44:30Z) - CktGNN: Circuit Graph Neural Network for Electronic Design Automation [67.29634073660239]
This paper presents a Circuit Graph Neural Network (CktGNN) that simultaneously automates the circuit topology generation and device sizing.
We introduce Open Circuit Benchmark (OCB), an open-sourced dataset that contains $10$K distinct operational amplifiers.
Our work paves the way toward a learning-based open-sourced design automation for analog circuits.
arXiv Detail & Related papers (2023-08-31T02:20:25Z) - NAR-Former V2: Rethinking Transformer for Universal Neural Network
Representation Learning [25.197394237526865]
We propose a modified Transformer-based universal neural network representation learning model NAR-Former V2.
Specifically, we take the network as a graph and design a straightforward tokenizer to encode the network into a sequence.
We incorporate the inductive representation learning capability of GNN into Transformer, enabling Transformer to generalize better when encountering unseen architecture.
arXiv Detail & Related papers (2023-06-19T09:11:04Z) - How neural networks learn to classify chaotic time series [77.34726150561087]
We study the inner workings of neural networks trained to classify regular-versus-chaotic time series.
We find that the relation between input periodicity and activation periodicity is key for the performance of LKCNN models.
arXiv Detail & Related papers (2023-06-04T08:53:27Z) - Learning to Learn with Generative Models of Neural Network Checkpoints [71.06722933442956]
We construct a dataset of neural network checkpoints and train a generative model on the parameters.
We find that our approach successfully generates parameters for a wide range of loss prompts.
We apply our method to different neural network architectures and tasks in supervised and reinforcement learning.
arXiv Detail & Related papers (2022-09-26T17:59:58Z) - Invertible Neural Networks for Graph Prediction [22.140275054568985]
In this work, we address conditional generation using deep invertible neural networks.
We adopt an end-to-end training approach since our objective is to address prediction and generation in the forward and backward processes at once.
arXiv Detail & Related papers (2022-06-02T17:28:33Z) - Generalizable Cross-Graph Embedding for GNN-based Congestion Prediction [22.974348682859322]
We propose a framework that can directly learn embeddings for the given netlist to enhance the quality of our node features.
By combining the learned embedding on top of the netlist with the GNNs, our method improves prediction performance, generalizes to new circuit lines, and is efficient in training, potentially saving over $90 %$ of runtime.
arXiv Detail & Related papers (2021-11-10T20:56:29Z) - PredRNN: A Recurrent Neural Network for Spatiotemporal Predictive
Learning [109.84770951839289]
We present PredRNN, a new recurrent network for learning visual dynamics from historical context.
We show that our approach obtains highly competitive results on three standard datasets.
arXiv Detail & Related papers (2021-03-17T08:28:30Z) - Inductive Graph Neural Networks for Spatiotemporal Kriging [13.666589510218738]
We develop an inductive graph neural network model to recover data for unsampled sensors on a network/graph structure.
Empirical results on several real-worldtemporal datasets demonstrate the effectiveness of our model.
arXiv Detail & Related papers (2020-06-13T01:23:44Z) - Training End-to-End Analog Neural Networks with Equilibrium Propagation [64.0476282000118]
We introduce a principled method to train end-to-end analog neural networks by gradient descent.
We show mathematically that a class of analog neural networks (called nonlinear resistive networks) are energy-based models.
Our work can guide the development of a new generation of ultra-fast, compact and low-power neural networks supporting on-chip learning.
arXiv Detail & Related papers (2020-06-02T23:38:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.