| --- |
| license: apache-2.0 |
| base_model: |
| - Qwen/Qwen2.5-7B-Instruct |
| pipeline_tag: text-to-3d |
| datasets: |
| - FreedomIntelligence/BlendNet |
| metrics: |
| - code_eval |
| tags: |
| - code |
| - render |
| - CAD |
| - 3D |
| - Modeling |
| - LLM |
| - bpy |
| - Blender |
| --- |
| |
| # 🤖 BlenderLLM: Training Large Language Models for Computer-Aided Design with Self-improvement |
|
|
| **BlenderLLM** is built using **Qwen2.5-Coder-7B-Instruct** as the base model. It has been fine-tuned on the **BlendNet** training dataset and further optimized through **Self-improvement** techniques to achieve the best performance. |
|
|
| For more details, please visit our [GitHub repository](https://github.com/FreedomIntelligence/BlenderLLM) or refer to our [arXiv paper](https://www.arxiv.org/abs/2412.14203). |
|
|
| ## 📖 Citation |
| ```angular2 |
| @misc{du2024blenderllmtraininglargelanguage, |
| title={BlenderLLM: Training Large Language Models for Computer-Aided Design with Self-improvement}, |
| author={Yuhao Du and Shunian Chen and Wenbo Zan and Peizhao Li and Mingxuan Wang and Dingjie Song and Bo Li and Yan Hu and Benyou Wang}, |
| year={2024}, |
| eprint={2412.14203}, |
| archivePrefix={arXiv}, |
| primaryClass={cs.HC}, |
| url={https://arxiv.org/abs/2412.14203}, |
| } |
| ``` |
|
|
| We are from the School of Data Science (SDS), the Chinese University of Hong Kong, Shenzhen (CUHKSZ). |
|
|