Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

QiMeng-CodeV-R1: Reasoning-Enhanced Verilog Generation

About

Large language models (LLMs) trained via reinforcement learning with verifiable reward (RLVR) have achieved breakthroughs on tasks with explicit, automatable verification, such as software programming and mathematical problems. Extending RLVR to electronic design automation (EDA), especially automatically generating hardware description languages (HDLs) like Verilog from natural-language (NL) specifications, however, poses three key challenges: the lack of automated and accurate verification environments, the scarcity of high-quality NL-code pairs, and the prohibitive computation cost of RLVR. To this end, we introduce CodeV-R1, an RLVR framework for training Verilog generation LLMs. First, we develop a rule-based testbench generator that performs robust equivalence checking against golden references. Second, we propose a round-trip data synthesis method that pairs open-source Verilog snippets with LLM-generated NL descriptions, verifies code-NL-code consistency via the generated testbench, and filters out inequivalent examples to yield a high-quality dataset. Third, we employ a two-stage "distill-then-RL" training pipeline: distillation for the cold start of reasoning abilities, followed by adaptive DAPO, our novel RLVR algorithm that can reduce training cost by adaptively adjusting sampling rate. The resulting model, CodeV-R1-7B, achieves 68.6% and 72.9% pass@1 on VerilogEval v2 and RTLLM v1.1, respectively, surpassing prior state-of-the-art by 12~20%, while even exceeding the performance of 671B DeepSeek-R1 on RTLLM. We have released our model, training code, and dataset to facilitate research in EDA and LLM communities.

Yaoyu Zhu, Di Huang, Hanqi Lyu, Xiaoyun Zhang, Chongxiao Li, Wenxuan Shi, Yutong Wu, Jianan Mu, Jinghua Wang, Yang Zhao, Pengwei Jin, Shuyao Cheng, Shengwen Liang, Xishan Zhang, Rui Zhang, Zidong Du, Qi Guo, Xing Hu, Yunji Chen• 2025

Related benchmarks

TaskDatasetResultRank
Verilog Code GenerationVerilogEval v1 (Human)
Pass@169.9
54
Verilog Code GenerationVerilogEval SR v2
Pass@168.8
34
Verilog Code GenerationVerilogEval CC v2
Pass@169.9
33
RTL generationVerilogEval 156 cases (test)
Pass@10.52
32
Hardware Verification Testbench GenerationCVDP-ECov
Coverage Pass Rate29.9
28
Verilog Code GenerationVerilogEval Machine v1
Pass@176.5
17
Verilog Code GenerationRTLLM v1
Pass@172.9
16
RTL generationRTLLM 50 cases (test)
Pass@10.32
16
Verilog Code GenerationRTLLM v2
Pass@168
13
Verilog GenerationRTLLM v2 (test)
Pass@168
11
Showing 10 of 12 rows

Other info

Follow for update