Natural Language to Verilog: Design of a Recurrent Spiking Neural Network using Large Language Models and ChatGPT
- URL: http://arxiv.org/abs/2405.01419v1
- Date: Thu, 2 May 2024 16:08:08 GMT
- Title: Natural Language to Verilog: Design of a Recurrent Spiking Neural Network using Large Language Models and ChatGPT
- Authors: Paola Vitolo, George Psaltakis, Michael Tomlinson, Gian Domenico Licciardo, Andreas G. Andreou,
- Abstract summary: We employ OpenAI's ChatGPT4 and natural language prompts to synthesize a RTL Verilog module of a programmable recurrent spiking neural network.
The resultant design was validated in three case studies, the exclusive OR, the IRIS flower classification and the MNIST hand-written digit classification, achieving accuracies of up to 96.6%.
- Score: 0.08388591755871733
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper investigates the use of Large Language Models (LLMs) for automating the generation of hardware description code, aiming to explore their potential in supporting and enhancing the development of efficient neuromorphic computing architectures. Building on our prior work, we employ OpenAI's ChatGPT4 and natural language prompts to synthesize a RTL Verilog module of a programmable recurrent spiking neural network, while also generating test benches to assess the system's correctness. The resultant design was validated in three case studies, the exclusive OR,the IRIS flower classification and the MNIST hand-written digit classification, achieving accuracies of up to 96.6%. To verify its synthesizability and implementability, the design was prototyped on a field-programmable gate array and implemented on SkyWater 130 nm technology by using an open-source electronic design automation flow. Additionally, we have submitted it to Tiny Tapeout 6 chip fabrication program to further evaluate the system on-chip performance in the future.
Related papers
- MATADOR: Automated System-on-Chip Tsetlin Machine Design Generation for Edge Applications [0.2663045001864042]
This paper presents MATADOR, an automated-to-silicon tool with GUI interface capable of optimized accelerator design for inference at the edge.
It offers automation of the full development pipeline: model training, system level design generation, design verification and deployment.
MATADOR accelerator designs are shown to be up to 13.4x faster, up to 7x more resource frugal and up to 2x more power efficient when compared to state-of-the-art Quantized and Binary Deep Neural Network implementations.
arXiv Detail & Related papers (2024-03-03T10:31:46Z) - Designing Silicon Brains using LLM: Leveraging ChatGPT for Automated
Description of a Spiking Neuron Array [1.137846619087643]
We present the prompts used to guide ChatGPT4 to produce a synthesizable and functional verilog description for a programmable Spiking Neuron Array ASIC.
This design flow showcases the current state of using ChatGPT4 for natural language driven hardware design.
arXiv Detail & Related papers (2024-01-25T21:21:38Z) - Zero-Shot RTL Code Generation with Attention Sink Augmented Large
Language Models [0.0]
This paper discusses the possibility of exploiting large language models to streamline the code generation process in hardware design.
The ability to use large language models on RTL code generation not only expedites design cycles but also facilitates the exploration of design spaces.
arXiv Detail & Related papers (2024-01-12T17:41:38Z) - Experimenting a New Programming Practice with LLMs [6.8035637735756715]
We develop a prototype named AISD (AI-aided Software Development)
It is capable of taking high-level (potentially vague) user requirements as inputs.
It generates detailed use cases, prototype system designs, and subsequently system implementation.
arXiv Detail & Related papers (2024-01-02T06:50:20Z) - Leveraging High-Level Synthesis and Large Language Models to Generate,
Simulate, and Deploy a Uniform Random Number Generator Hardware Design [0.0]
We present a new high-level synthesis methodology for using large language model tools to generate hardware designs.
As a case study, we use our methodology to generate a permuted congruential random number generator design with a wishbone interface.
arXiv Detail & Related papers (2023-11-06T19:58:26Z) - Generative Input: Towards Next-Generation Input Methods Paradigm [49.98958865125018]
We propose a novel Generative Input paradigm named GeneInput.
It uses prompts to handle all input scenarios and other intelligent auxiliary input functions, optimizing the model with user feedback to deliver personalized results.
The results demonstrate that we have achieved state-of-the-art performance for the first time in the Full-mode Key-sequence to Characters(FK2C) task.
arXiv Detail & Related papers (2023-11-02T12:01:29Z) - CktGNN: Circuit Graph Neural Network for Electronic Design Automation [67.29634073660239]
This paper presents a Circuit Graph Neural Network (CktGNN) that simultaneously automates the circuit topology generation and device sizing.
We introduce Open Circuit Benchmark (OCB), an open-sourced dataset that contains $10$K distinct operational amplifiers.
Our work paves the way toward a learning-based open-sourced design automation for analog circuits.
arXiv Detail & Related papers (2023-08-31T02:20:25Z) - PDSketch: Integrated Planning Domain Programming and Learning [86.07442931141637]
We present a new domain definition language, named PDSketch.
It allows users to flexibly define high-level structures in the transition models.
Details of the transition model will be filled in by trainable neural networks.
arXiv Detail & Related papers (2023-03-09T18:54:12Z) - Decentralizing Feature Extraction with Quantum Convolutional Neural
Network for Automatic Speech Recognition [101.69873988328808]
We build upon a quantum convolutional neural network (QCNN) composed of a quantum circuit encoder for feature extraction.
An input speech is first up-streamed to a quantum computing server to extract Mel-spectrogram.
The corresponding convolutional features are encoded using a quantum circuit algorithm with random parameters.
The encoded features are then down-streamed to the local RNN model for the final recognition.
arXiv Detail & Related papers (2020-10-26T03:36:01Z) - Exploring Software Naturalness through Neural Language Models [56.1315223210742]
The Software Naturalness hypothesis argues that programming languages can be understood through the same techniques used in natural language processing.
We explore this hypothesis through the use of a pre-trained transformer-based language model to perform code analysis tasks.
arXiv Detail & Related papers (2020-06-22T21:56:14Z) - Synthetic Datasets for Neural Program Synthesis [66.20924952964117]
We propose a new methodology for controlling and evaluating the bias of synthetic data distributions over both programs and specifications.
We demonstrate, using the Karel DSL and a small Calculator DSL, that training deep networks on these distributions leads to improved cross-distribution generalization performance.
arXiv Detail & Related papers (2019-12-27T21:28:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.