Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

HPC-Coder-V2: Studying Code LLMs Across Low-Resource Parallel Languages

About

Large Language Model (LLM) based coding tools have been tremendously successful as software development assistants, yet they are often designed for general purpose programming tasks and perform poorly for more specialized domains such as high performance computing. Creating specialized models and tools for these domains is crucial towards gaining the benefits of LLMs in areas such as HPC. While previous work has explored HPC-specific models, LLMs still struggle to generate parallel code and it is not at all clear what hurdles are still holding back these LLMs and what must be done to overcome them. In this work, we conduct an in-depth study along the many axes of fine-tuning a specialized HPC LLM in order to better understand the challenges. Based on our findings we fine-tune and evaluate a specialized HPC LLM that is shown to be the best performing open-source code LLM for parallel code generation to date.

Aman Chaturvedi, Daniel Nichols, Siddharth Singh, Abhinav Bhatele• 2024

Related benchmarks

TaskDatasetResultRank
Dot ProductCUDA-LLM task suite
Execution Time (ms)7.36
9
Matrix CopyCUDA-LLM task suite
Execution Time8.98
9
Matrix MultiplicationCUDA-LLM kernels task suite 1.0 (test)
Execution Time7.88
9
Matrix MultiplicationCUDA-LLM kernels task suite Matrix Multiplication
Execution Time (s)7.88
9
Matrix TransposeCUDA-LLM task suite
Time9.88
9
Mean Square ErrorCUDA-LLM kernels task suite 1.0 (test)
Execution Time8.78
9
ReductionCUDA-LLM task suite
Time10.18
9
ReLU Activation FunctionCUDA-LLM task suite
Time8.02
9
Reverse ArrayCUDA-LLM task suite
Execution Time7.28
9
CUDA Kernel OptimizationParEval 5-task subset
Pass@120
9
Showing 10 of 10 rows

Other info

Follow for update