FALCON: An ML Framework for Fully Automated Layout-Constrained Analog Circuit Design
- URL: http://arxiv.org/abs/2505.21923v2
- Date: Mon, 27 Oct 2025 22:42:49 GMT
- Title: FALCON: An ML Framework for Fully Automated Layout-Constrained Analog Circuit Design
- Authors: Asal Mehradfar, Xuzhe Zhao, Yilun Huang, Emir Ceyani, Yankai Yang, Shihao Han, Hamidreza Aghasi, Salman Avestimehr,
- Abstract summary: We introduce FALCON, a unified machine learning framework that enables specification-driven analog circuit synthesis.<n>FALCON first selects an appropriate circuit topology using a performance-driven classifier guided by human designs.<n>Next, it employs a custom, edge-centric graph neural network trained to map circuit topology and parameters to performance.<n>This inference is guided by a differentiable layout cost, derived from analytical equations parasitic and frequency-dependent effects, and constrained by design rules.
- Score: 15.502124836790438
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Designing analog circuits from performance specifications is a complex, multi-stage process encompassing topology selection, parameter inference, and layout feasibility. We introduce FALCON, a unified machine learning framework that enables fully automated, specification-driven analog circuit synthesis through topology selection and layout-constrained optimization. Given a target performance, FALCON first selects an appropriate circuit topology using a performance-driven classifier guided by human design heuristics. Next, it employs a custom, edge-centric graph neural network trained to map circuit topology and parameters to performance, enabling gradient-based parameter inference through the learned forward model. This inference is guided by a differentiable layout cost, derived from analytical equations capturing parasitic and frequency-dependent effects, and constrained by design rules. We train and evaluate FALCON on a large-scale custom dataset of 1M analog mm-wave circuits, generated and simulated using Cadence Spectre across 20 expert-designed topologies. Through this evaluation, FALCON demonstrates >99% accuracy in topology inference, <10% relative error in performance prediction, and efficient layout-aware design that completes in under 1 second per instance. Together, these results position FALCON as a practical and extensible foundation model for end-to-end analog circuit design automation.
Related papers
- Artificial Intelligence-Driven Network-on-Chip Design Space Exploration: Neural Network Architectures for Design [0.0]
Network-on-Chip (NoC) design requires exploring a high-dimensional configuration space to satisfy stringent throughput requirements and latency constraints.<n>Traditional design space exploration techniques are often slow and struggle to handle complex, non-linear parameter interactions.<n>This work presents a machine learning-driven framework that automates NoC design space exploration using BookSim simulations and reverse neural network models.
arXiv Detail & Related papers (2025-11-27T13:00:22Z) - ZeroSim: Zero-Shot Analog Circuit Evaluation with Unified Transformer Embeddings [36.00558662624576]
ZeroSim is a transformer-based performance modeling framework designed to achieve robust in-distribution generalization across trained topologies.<n>We show that ZeroSim significantly outperforms baseline models such as multilayer perceptrons, graph neural networks and transformers.<n>When integrated into a reinforcement learning-based parameter optimization pipeline, ZeroSim achieves a remarkable speedup (13x) compared to conventional SPICE simulations.
arXiv Detail & Related papers (2025-11-10T22:08:30Z) - URDF-Anything: Constructing Articulated Objects with 3D Multimodal Language Model [76.08429266631823]
We propose an end-to-end automatic reconstruction framework based on a 3D multimodal large language model (MLLM)<n>URDF-Anything utilizes an autoregressive prediction framework based on point-cloud and text multimodal input to jointly optimize geometric segmentation and kinematic parameter prediction.<n> Experiments on both simulated and real-world datasets demonstrate that our method significantly outperforms existing approaches.
arXiv Detail & Related papers (2025-11-02T13:45:51Z) - Learning from Scratch: Structurally-masked Transformer for Next Generation Lib-free Simulation [3.7467132954493536]
This paper proposes a neural framework for power and timing prediction of multi-stage data path.<n>To the best of our knowledge, this is the first language-based, netlist-aware neural network designed explicitly for standard cells.
arXiv Detail & Related papers (2025-07-23T10:46:25Z) - AnalogXpert: Automating Analog Topology Synthesis by Incorporating Circuit Design Expertise into Large Language Models [10.986618360243526]
We propose AnalogXpert, a LLM-based agent aiming at solving practical topology synthesis problem.<n>First, we represent analog topology as SPICE code and introduce a subcircuit library to reduce the design space.<n>Second, we decompose the problem into two sub-tasks through the use of CoT and incontext learning techniques.<n>Third, we introduce a proofreading strategy that allows LLMs to incrementally correct the errors in the initial design.
arXiv Detail & Related papers (2024-12-17T09:08:08Z) - MATADOR: Automated System-on-Chip Tsetlin Machine Design Generation for Edge Applications [0.2663045001864042]
This paper presents MATADOR, an automated-to-silicon tool with GUI interface capable of optimized accelerator design for inference at the edge.
It offers automation of the full development pipeline: model training, system level design generation, design verification and deployment.
MATADOR accelerator designs are shown to be up to 13.4x faster, up to 7x more resource frugal and up to 2x more power efficient when compared to state-of-the-art Quantized and Binary Deep Neural Network implementations.
arXiv Detail & Related papers (2024-03-03T10:31:46Z) - Machine Learning Driven Global Optimisation Framework for Analog Circuit Design [0.0]
We propose a machine learning-driven optimisation framework for analog circuit design.
We employ machine learning models and spice simulations to direct the optimisation algorithm.
arXiv Detail & Related papers (2024-02-27T03:51:00Z) - CktGNN: Circuit Graph Neural Network for Electronic Design Automation [67.29634073660239]
This paper presents a Circuit Graph Neural Network (CktGNN) that simultaneously automates the circuit topology generation and device sizing.
We introduce Open Circuit Benchmark (OCB), an open-sourced dataset that contains $10$K distinct operational amplifiers.
Our work paves the way toward a learning-based open-sourced design automation for analog circuits.
arXiv Detail & Related papers (2023-08-31T02:20:25Z) - Learning to Design Analog Circuits to Meet Threshold Specifications [2.5818330243826924]
We propose a method for generating from simulation data a dataset on which a system can be trained to design circuits to meet threshold specifications.
We show that our method consistently reaches success rate better than 90% at 5% error margin, while also improving data efficiency by upward of an order of magnitude.
arXiv Detail & Related papers (2023-07-25T23:25:05Z) - End-to-End Meta-Bayesian Optimisation with Transformer Neural Processes [52.818579746354665]
This paper proposes the first end-to-end differentiable meta-BO framework that generalises neural processes to learn acquisition functions via transformer architectures.
We enable this end-to-end framework with reinforcement learning (RL) to tackle the lack of labelled acquisition data.
arXiv Detail & Related papers (2023-05-25T10:58:46Z) - Efficient Automatic Machine Learning via Design Graphs [72.85976749396745]
We propose FALCON, an efficient sample-based method to search for the optimal model design.
FALCON features 1) a task-agnostic module, which performs message passing on the design graph via a Graph Neural Network (GNN), and 2) a task-specific module, which conducts label propagation of the known model performance information.
We empirically show that FALCON can efficiently obtain the well-performing designs for each task using only 30 explored nodes.
arXiv Detail & Related papers (2022-10-21T21:25:59Z) - Squeezeformer: An Efficient Transformer for Automatic Speech Recognition [99.349598600887]
Conformer is the de facto backbone model for various downstream speech tasks based on its hybrid attention-convolution architecture.
We propose the Squeezeformer model, which consistently outperforms the state-of-the-art ASR models under the same training schemes.
arXiv Detail & Related papers (2022-06-02T06:06:29Z) - Pretraining Graph Neural Networks for few-shot Analog Circuit Modeling
and Design [68.1682448368636]
We present a supervised pretraining approach to learn circuit representations that can be adapted to new unseen topologies or unseen prediction tasks.
To cope with the variable topological structure of different circuits we describe each circuit as a graph and use graph neural networks (GNNs) to learn node embeddings.
We show that pretraining GNNs on prediction of output node voltages can encourage learning representations that can be adapted to new unseen topologies or prediction of new circuit level properties.
arXiv Detail & Related papers (2022-03-29T21:18:47Z) - DAIS: Automatic Channel Pruning via Differentiable Annealing Indicator
Search [55.164053971213576]
convolutional neural network has achieved great success in fulfilling computer vision tasks despite large computation overhead.
Structured (channel) pruning is usually applied to reduce the model redundancy while preserving the network structure.
Existing structured pruning methods require hand-crafted rules which may lead to tremendous pruning space.
arXiv Detail & Related papers (2020-11-04T07:43:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.