Add paper link and improve model card metadata

#2
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +22 -3
README.md CHANGED
@@ -1,8 +1,9 @@
1
  ---
2
- license: apache-2.0
3
  language:
4
  - code
5
  library_name: transformers
 
 
6
  tags:
7
  - masked-diffusion
8
  - code-generation
@@ -11,7 +12,12 @@ tags:
11
 
12
  ## Open Diffusion Large Language Models for Code Generation
13
 
14
- This repository contains the weights and custom code for the **fredzzp/open-dcoder-0.5B** model, a masked diffusion model for code generation based on the Qwen2 architecture.
 
 
 
 
 
15
 
16
  This model uses bidirectional attention and must be used with the custom `diffusion_generate` method.
17
 
@@ -23,8 +29,8 @@ First, make sure you have the latest `transformers` library installed.
23
  pip install transformers torch huggingface_hub
24
  ```
25
 
 
26
 
27
- You can then use the model for generation. Note: You must pass trust_remote_code=True to load the custom model architecture.
28
  ```python
29
  import torch
30
  from transformers import AutoTokenizer, AutoModelForCausalLM
@@ -59,3 +65,16 @@ generated_text = tokenizer.decode(outputs.sequences[0][prompt_len:], skip_specia
59
  print("--- Generated Code ---")
60
  print(generated_text)
61
  ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
 
2
  language:
3
  - code
4
  library_name: transformers
5
+ license: apache-2.0
6
+ pipeline_tag: text-generation
7
  tags:
8
  - masked-diffusion
9
  - code-generation
 
12
 
13
  ## Open Diffusion Large Language Models for Code Generation
14
 
15
+ This repository contains the weights and custom code for the **fredzzp/open-dcoder-0.5B** model, a masked diffusion model for code generation based on the Qwen2 architecture.
16
+
17
+ The model was introduced in the paper [Don't Retrain, Align: Adapting Autoregressive LMs to Diffusion LMs via Representation Alignment](https://huggingface.co/papers/2605.06885).
18
+
19
+ - **Code:** [pengzhangzhi/Open-dLLM](https://github.com/pengzhangzhi/Open-dLLM)
20
+ - **Blog:** [Notion Blog](https://oval-shell-31c.notion.site/Open-Diffusion-Large-Language-Model-25e03bf6136480b7a4ebe3d53be9f68a?pvs=74)
21
 
22
  This model uses bidirectional attention and must be used with the custom `diffusion_generate` method.
23
 
 
29
  pip install transformers torch huggingface_hub
30
  ```
31
 
32
+ You can then use the model for generation. Note: You must pass `trust_remote_code=True` to load the custom model architecture.
33
 
 
34
  ```python
35
  import torch
36
  from transformers import AutoTokenizer, AutoModelForCausalLM
 
65
  print("--- Generated Code ---")
66
  print(generated_text)
67
  ```
68
+
69
+ ## Citation
70
+
71
+ ```bibtex
72
+ @misc{opendllm2025,
73
+ title = {Open-dLLM: Open Diffusion Large Language Models},
74
+ author = {Fred Zhangzhi Peng, Shuibai Zhang, Alex Tong, and contributors},
75
+ year = {2025},
76
+ howpublished = {\url{https://github.com/pengzhangzhi/Open-dLLM}},
77
+ note = {Blog: \url{https://oval-shell-31c.notion.site/Open-Diffusion-Large-Language-Model-25e03bf6136480b7a4ebe3d53be9f68a?pvs=74},
78
+ Model: \url{https://huggingface.co/fredzzp/open-dcoder-0.5B}}
79
+ }
80
+ ```