suayptalha commited on
Commit
386ea89
·
verified ·
1 Parent(s): 946cd63

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +25 -8
README.md CHANGED
@@ -61,14 +61,31 @@ Despite being an intermediate checkpoint, the 1st-Stage model demonstrates highl
61
 
62
  ## 💻 Usage
63
 
64
- Inference requires generating text via a discrete diffusion process rather than causal next-token prediction. We recommend using the `dllm` library.
65
-
66
- **Recommended Generation Parameters:**
67
- * **Steps:** 64 to 128
68
- * **Temperature:** 0.1
69
- * **Block Length:** 32
70
- * **Repetition Penalty:** 1.2
71
- * **Remask Strategy:** `low_conf`
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
72
 
73
  ## ⚠️ Limitations
74
 
 
61
 
62
  ## 💻 Usage
63
 
64
+ Because Diffutron is a Masked Diffusion Language Model, it requires inference strategies distinct from standard causal generation. We recommend using the `dllm` library or custom generation loops tailored for discrete diffusion.
65
+
66
+ ### 1. Install the dllm Library:
67
+ ```bash
68
+ git clone https://github.com/Diffutron/dllm.git
69
+ cd dllm
70
+ pip install -e .
71
+ ```
72
+ ### 2. Chat via Interaction Mode:
73
+
74
+ ```bash
75
+ python -u examples/bert/chat.py \
76
+ --model_name_or_path "diffutron/DiffutronLM-0.3B-1st-Stage" \
77
+ --chat True \
78
+ --steps 64 \
79
+ --max_new_tokens 64 \
80
+ --temperature 0.1 \
81
+ --block_length 32 \
82
+ --repetition_penalty 1.2 \
83
+ --remasking "low_confidence" \
84
+ --stochastic_transfer False \
85
+ --cfg_scale 0.0
86
+ ```
87
+
88
+ For other inference modes, see [dllm](https://github.com/Diffutron/dllm) library.
89
 
90
  ## ⚠️ Limitations
91