Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

BrepCoder: A Unified Multimodal Large Language Model for Multi-task B-rep Reasoning

About

Recent advancements in deep learning have actively addressed complex challenges within the Computer-Aided Design (CAD) domain.However, most existing approaches rely on task-specifi c models requiring structural modifi cations for new tasks, and they predominantly focus on point clouds or images rather than the industry-standard Boundary Representation (B-rep) format. To address these limitations, we propose BrepCoder, a unifi ed Multimodal Large Language Model (MLLM) that performs diverse CAD tasks from B-rep inputs. By leveraging the code generation capabilities of Large Language Models (LLMs), we convert CAD modeling sequences into Python-like code and align them with B-rep. We then adopt a two-stage training strategy: First, pre-training on reverse engineering to learn geometric features and design logic. Second, eff ectively extending the model to various downstream tasks such as completion, error correction, and CAD-QA. Consequently, by interpreting B-rep as structural code, BrepCoder achieves superior generalization across diverse tasks, demonstrating its potential as a general-purpose CAD agent.

Mingi Kim, Yongjun Kim, Jungwoo Kang, Hyungki Kim• 2026

Related benchmarks

TaskDatasetResultRank
CAD CompletionDeepCAD
Command Accuracy (ACC)92.69
6
Reverse EngineeringDeepCAD
Median Chamfer Distance (Med. CD)0.464
6
Error CorrectionError Correction
Median CD0.335
4
Showing 3 of 3 rows

Other info

Follow for update