UPDF AI

RTLCoder: Outperforming GPT-3.5 in Design RTL Generation with Our Open-Source Dataset and Lightweight Solution

Shang Liu,Wenji Fang,3 Authors,Zhiyao Xie

2024 · DOI: 10.1109/LAD62341.2024.10691788
67 Citations

TLDR

This study presents a new customized LLM solution with a modest parameter count, achieving better performance than GPT-3.5 on all representative benchmarks for RTL code generation and outperforms GPT-4 in VerilogEval Machine benchmark.

Abstract

The automatic generation of RTL code (e.g., Verilog) using natural language instructions and large language models (LLMs) has attracted significant research interest recently. However, most existing approaches heavily rely on commercial LLMs such as ChatGPT, while open-source LLMs tailored for this specific design generation task exhibit notably inferior performance. The absence of high-quality open-source solutions restricts the flexibility and data privacy of this emerging technique. In this study, we present a new customized LLM solution with a modest parameter count of only 7B, achieving better performance than GPT-3.5 on all representative benchmarks for RTL code generation. Especially, it outperforms GPT-4 in VerilogEval Machine benchmark. This remarkable balance between accuracy and efficiency is made possible by leveraging our new RTL code dataset and a customized LLM algorithm, both of which have been made fully open-source1.