Add metadata and improve model card
#1
by nielsr HF Staff - opened
README.md
CHANGED
|
@@ -1,19 +1,23 @@
|
|
| 1 |
---
|
| 2 |
language:
|
| 3 |
- en
|
| 4 |
-
.
|
|
|
|
|
|
|
| 5 |
---
|
| 6 |
|
| 7 |
-
# AscendKernelGen/KernelGen-LM-4B
|
| 8 |
|
| 9 |

|
| 10 |
-
[](https://
|
| 11 |
|
| 12 |
KernelGen-LM-4B is a state-of-the-art domain-adaptive large language model specialized for low-level NPU kernel generation, specifically for the Huawei Ascend architecture using the AscendC programming language. Built upon the Qwen3-4B backbone, it is trained on the Ascend-CoT dataset and refined via reinforcement learning with execution feedback.
|
| 13 |
|
|
|
|
|
|
|
| 14 |
**Other artifacts:**
|
| 15 |
-
*
|
| 16 |
-
*
|
| 17 |
|
| 18 |
## Introduction
|
| 19 |
|
|
@@ -25,10 +29,13 @@ Our framework, **AscendKernelGen (AKGen)**, bridges the gap between general-purp
|
|
| 25 |
* **Performance:** The model demonstrates siginificant improvement on complex Level-2 kernels compared to baselines, and effectively solving tasks where general-purpose models (like Qwen3, Llama3.1) fail completely.
|
| 26 |
|
| 27 |
## Citation
|
|
|
|
|
|
|
| 28 |
@article{cao2026ascendkernelgen,
|
| 29 |
title={AscendKernelGen: A Systematic Study of LLM-Based Kernel Generation for Neural Processing Units},
|
| 30 |
author={Xinzi Cao and Jianyang Zhai and Pengfei Li and Zhiheng Hu and Cen Yan and Bingxu Mu and Guanghuan Fang and Bin She and Jiayu Li and Yihan Su and Dongyang Tao and Xiansong Huang and Fan Xu and Feidiao Yang and Yao Lu and Chang-Dong Wang and Yutong Lu and Weicheng Xue and Bin Zhou and Yonghong Tian},
|
| 31 |
journal={arXiv preprint arXiv:2601.07160},
|
| 32 |
year={2026},
|
| 33 |
-
url=https://arxiv.org/abs/2601.07160
|
| 34 |
-
}
|
|
|
|
|
|
| 1 |
---
|
| 2 |
language:
|
| 3 |
- en
|
| 4 |
+
license: apache-2.0
|
| 5 |
+
library_name: transformers
|
| 6 |
+
pipeline_tag: text-generation
|
| 7 |
---
|
| 8 |
|
| 9 |
+
# AscendKernelGen / KernelGen-LM-4B
|
| 10 |
|
| 11 |

|
| 12 |
+
[](https://huggingface.co/papers/2601.07160)
|
| 13 |
|
| 14 |
KernelGen-LM-4B is a state-of-the-art domain-adaptive large language model specialized for low-level NPU kernel generation, specifically for the Huawei Ascend architecture using the AscendC programming language. Built upon the Qwen3-4B backbone, it is trained on the Ascend-CoT dataset and refined via reinforcement learning with execution feedback.
|
| 15 |
|
| 16 |
+
This model was introduced in the paper [AscendKernelGen: A Systematic Study of LLM-Based Kernel Generation for Neural Processing Units](https://huggingface.co/papers/2601.07160).
|
| 17 |
+
|
| 18 |
**Other artifacts:**
|
| 19 |
+
* **GitHub Repository:** [NPUKernelBench](https://github.com/weich97/NPUKernelBench)
|
| 20 |
+
* **Evaluation Framework (OpenI):** [NPUKernelBench](https://git.openi.org.cn/PCL-Benchmark/NPUKernelBench)
|
| 21 |
|
| 22 |
## Introduction
|
| 23 |
|
|
|
|
| 29 |
* **Performance:** The model demonstrates siginificant improvement on complex Level-2 kernels compared to baselines, and effectively solving tasks where general-purpose models (like Qwen3, Llama3.1) fail completely.
|
| 30 |
|
| 31 |
## Citation
|
| 32 |
+
|
| 33 |
+
```bibtex
|
| 34 |
@article{cao2026ascendkernelgen,
|
| 35 |
title={AscendKernelGen: A Systematic Study of LLM-Based Kernel Generation for Neural Processing Units},
|
| 36 |
author={Xinzi Cao and Jianyang Zhai and Pengfei Li and Zhiheng Hu and Cen Yan and Bingxu Mu and Guanghuan Fang and Bin She and Jiayu Li and Yihan Su and Dongyang Tao and Xiansong Huang and Fan Xu and Feidiao Yang and Yao Lu and Chang-Dong Wang and Yutong Lu and Weicheng Xue and Bin Zhou and Yonghong Tian},
|
| 37 |
journal={arXiv preprint arXiv:2601.07160},
|
| 38 |
year={2026},
|
| 39 |
+
url={https://arxiv.org/abs/2601.07160}
|
| 40 |
+
}
|
| 41 |
+
```
|