ZeroSim: Zero-Shot Analog Circuit Evaluation with Unified Transformer Embeddings
- URL: http://arxiv.org/abs/2511.07658v1
- Date: Wed, 12 Nov 2025 01:09:54 GMT
- Title: ZeroSim: Zero-Shot Analog Circuit Evaluation with Unified Transformer Embeddings
- Authors: Xiaomeng Yang, Jian Gao, Yanzhi Wang, Xuan Zhang,
- Abstract summary: ZeroSim is a transformer-based performance modeling framework designed to achieve robust in-distribution generalization across trained topologies.<n>We show that ZeroSim significantly outperforms baseline models such as multilayer perceptrons, graph neural networks and transformers.<n>When integrated into a reinforcement learning-based parameter optimization pipeline, ZeroSim achieves a remarkable speedup (13x) compared to conventional SPICE simulations.
- Score: 36.00558662624576
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Although recent advancements in learning-based analog circuit design automation have tackled tasks such as topology generation, device sizing, and layout synthesis, efficient performance evaluation remains a major bottleneck. Traditional SPICE simulations are time-consuming, while existing machine learning methods often require topology-specific retraining or manual substructure segmentation for fine-tuning, hindering scalability and adaptability. In this work, we propose ZeroSim, a transformer-based performance modeling framework designed to achieve robust in-distribution generalization across trained topologies under novel parameter configurations and zero-shot generalization to unseen topologies without any fine-tuning. We apply three key enabling strategies: (1) a diverse training corpus of 3.6 million instances covering over 60 amplifier topologies, (2) unified topology embeddings leveraging global-aware tokens and hierarchical attention to robustly generalize to novel circuits, and (3) a topology-conditioned parameter mapping approach that maintains consistent structural representations independent of parameter variations. Our experimental results demonstrate that ZeroSim significantly outperforms baseline models such as multilayer perceptrons, graph neural networks and transformers, delivering accurate zero-shot predictions across different amplifier topologies. Additionally, when integrated into a reinforcement learning-based parameter optimization pipeline, ZeroSim achieves a remarkable speedup (13x) compared to conventional SPICE simulations, underscoring its practical value for a wide range of analog circuit design automation tasks.
Related papers
- AstRL: Analog and Mixed-Signal Circuit Synthesis with Deep Reinforcement Learning [0.8633481456093186]
This work casts circuit design as a graph generation problem and introduces a novel method of AMS synthesis driven by deep reinforcement learning (AstRL)<n>AstRL generates circuits directly optimized for user-specified targets within a simulator-embedded environment that provides ground-truth feedback during training.<n> Experimental results for three realistic design tasks illustrate substantial improvements in conventional design metrics over state-of-the-art baselines.
arXiv Detail & Related papers (2026-02-12T20:52:39Z) - URDF-Anything: Constructing Articulated Objects with 3D Multimodal Language Model [76.08429266631823]
We propose an end-to-end automatic reconstruction framework based on a 3D multimodal large language model (MLLM)<n>URDF-Anything utilizes an autoregressive prediction framework based on point-cloud and text multimodal input to jointly optimize geometric segmentation and kinematic parameter prediction.<n> Experiments on both simulated and real-world datasets demonstrate that our method significantly outperforms existing approaches.
arXiv Detail & Related papers (2025-11-02T13:45:51Z) - FALCON: An ML Framework for Fully Automated Layout-Constrained Analog Circuit Design [15.502124836790438]
We introduce FALCON, a unified machine learning framework that enables specification-driven analog circuit synthesis.<n>FALCON first selects an appropriate circuit topology using a performance-driven classifier guided by human designs.<n>Next, it employs a custom, edge-centric graph neural network trained to map circuit topology and parameters to performance.<n>This inference is guided by a differentiable layout cost, derived from analytical equations parasitic and frequency-dependent effects, and constrained by design rules.
arXiv Detail & Related papers (2025-05-28T03:16:08Z) - Fusing Global and Local: Transformer-CNN Synergy for Next-Gen Current Estimation [4.945568106952893]
This paper presents a hybrid model combining Transformer and CNN for predicting the current waveform in signal lines.<n>It replaces the complex Newton iteration process used in traditional SPICE simulations, leveraging the powerful sequence modeling capabilities of the Transformer framework.<n> Experimental results demonstrate that, compared to traditional SPICE simulations, the proposed algorithm achieves an error of only 0.0098.
arXiv Detail & Related papers (2025-04-08T19:42:10Z) - Supervised Learning for Analog and RF Circuit Design: Benchmarks and Comparative Insights [10.354863964933019]
This study explores supervised ML-based approaches for designing circuit parameters from performance specifications across various circuit types.<n>Our results show that simpler circuits, such as low-noise amplifiers, achieve exceptional accuracy with mean relative errors as low as 0.3%.<n>For heterogeneous circuits, our approach achieves an 88% reduction in errors with increased training data, with the receiver achieving a mean relative error as low as 0.23%.
arXiv Detail & Related papers (2025-01-21T02:48:23Z) - Auto-Train-Once: Controller Network Guided Automatic Network Pruning from Scratch [72.26822499434446]
Auto-Train-Once (ATO) is an innovative network pruning algorithm designed to automatically reduce the computational and storage costs of DNNs.
We provide a comprehensive convergence analysis as well as extensive experiments, and the results show that our approach achieves state-of-the-art performance across various model architectures.
arXiv Detail & Related papers (2024-03-21T02:33:37Z) - Machine Learning Driven Global Optimisation Framework for Analog Circuit Design [0.0]
We propose a machine learning-driven optimisation framework for analog circuit design.
We employ machine learning models and spice simulations to direct the optimisation algorithm.
arXiv Detail & Related papers (2024-02-27T03:51:00Z) - Transformers as Statisticians: Provable In-Context Learning with
In-Context Algorithm Selection [88.23337313766353]
This work first provides a comprehensive statistical theory for transformers to perform ICL.
We show that transformers can implement a broad class of standard machine learning algorithms in context.
A emphsingle transformer can adaptively select different base ICL algorithms.
arXiv Detail & Related papers (2023-06-07T17:59:31Z) - Pretraining Graph Neural Networks for few-shot Analog Circuit Modeling
and Design [68.1682448368636]
We present a supervised pretraining approach to learn circuit representations that can be adapted to new unseen topologies or unseen prediction tasks.
To cope with the variable topological structure of different circuits we describe each circuit as a graph and use graph neural networks (GNNs) to learn node embeddings.
We show that pretraining GNNs on prediction of output node voltages can encourage learning representations that can be adapted to new unseen topologies or prediction of new circuit level properties.
arXiv Detail & Related papers (2022-03-29T21:18:47Z) - Global Vision Transformer Pruning with Hessian-Aware Saliency [93.33895899995224]
This work challenges the common design philosophy of the Vision Transformer (ViT) model with uniform dimension across all the stacked blocks in a model stage.
We derive a novel Hessian-based structural pruning criteria comparable across all layers and structures, with latency-aware regularization for direct latency reduction.
Performing iterative pruning on the DeiT-Base model leads to a new architecture family called NViT (Novel ViT), with a novel parameter that utilizes parameters more efficiently.
arXiv Detail & Related papers (2021-10-10T18:04:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.