Generating CAD Code with Vision-Language Models for 3D Designs: |Paper|
1. CADCodeVerify approach: a novel approach to iteratively verify and improve the design output of 3D objects generated from CAD code.
2. CADPrompt dataset: the first benchmark for CAD code generation, consisting of 200 natural language prompts paired with expert-annotated scripting code for 3D objects to benchmark progress.
We introduce CADCodeVerify, a novel approach to iteratively verify and improve 3D objects generated from CAD code. Our approach works by producing ameliorative feedback by prompting a Vision-Language Model (VLM) to generate and answer a set of validation questions to verify the generated object and prompt the VLM to correct deviations.
We introduce a new benchmark, CADPrompt, featuring 200 3D objects represented in the Standard Triangle Language (STL) format. Each sample includes:
- The 3D object in STL format sourced from DeepCAD [1].
- A language-based description of the 3D object.
- The 3D object in OBJ format, detailing vertices, faces, and edges in standard 3D geometry.
- A
.jsonfile containing the CAD commands and their parameters. - The Python code that generates the 3D object, written by a CAD design expert.
The table below provides statistics for the CADPrompt dataset, including vertex and face counts of ground truth 3D objects, the lengths of language descriptions, and the corresponding Python code. The dataset is divided into simple and complex categories.
[1] Wu, Rundi, Chang Xiao, and Changxi Zheng. "Deepcad: A deep generative network for computer-aided design models." Proceedings of the IEEE/CVF International Conference on Computer Vision. 2021.
If this dataset contributes to your work, please consider citing the following publications.
@article{alrashedy2024generating,
title={Generating CAD Code with Vision-Language Models for 3D Designs},
author={Alrashedy, Kamel and Tambwekar, Pradyumna and Zaidi, Zulfiqar and Langwasser, Megan and Xu, Wei and Gombolay, Matthew},
booktitle={International conference on learning representations (ICLR)},
year={2025}
}
@inproceedings{wu2021deepcad,
title={Deepcad: A deep generative network for computer-aided design models},
author={Wu, Rundi and Xiao, Chang and Zheng, Changxi},
booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision},
pages={6772--6782},
year={2021}
}


