| --- |
| license: apache-2.0 |
| tags: |
| - diffusion |
| - dream |
| - gguf |
| - cpu-inference |
| - diffuse-cpp |
| language: |
| - en |
| base_model: Dream-org/Dream-v0-Instruct-7B |
| pipeline_tag: text-generation |
| --- |
| |
| # Dream-v0-Instruct-7B-GGUF |
|
|
| GGUF quantizations of [Dream-org/Dream-v0-Instruct-7B](https://huggingface.co/Dream-org/Dream-v0-Instruct-7B) for use with [diffuse-cpp](https://github.com/iafiscal1212/diffuse-cpp), the first C++ inference engine for Diffusion Language Models. |
|
|
| Dream is a masked diffusion language model based on the Qwen2.5-7B backbone with Grouped Query Attention (GQA). It generates all tokens in parallel through iterative refinement, excelling at math and factual tasks. |
|
|
| **Dream correctly solves 15 x 23 = 345 in just 2 denoising steps at 21.6 tok/s — 2.5x faster than llama.cpp.** |
|
|
| ## Available Quantizations |
|
|
| | File | Type | Size | Description | |
| |------|------|------|-------------| |
| | `dream-7b-f16.gguf` | F16 | ~15 GB | Full precision, best quality | |
| | `dream-7b-q8_0.gguf` | Q8_0 | ~8.2 GB | 8-bit quantization, near-lossless | |
| | `dream-7b-q4km.gguf` | Q4_K_M | ~5.0 GB | 4-bit mixed, best speed/quality ratio | |
| |
| **Recommended:** Q4_K_M for most users. |
| |
| ## Quick Start |
| |
| ```bash |
| # Download |
| huggingface-cli download diffuse-cpp/Dream-v0-Instruct-7B-GGUF dream-7b-q4km.gguf |
| |
| # Build diffuse-cpp (v0.2.0+) |
| git clone --recursive https://github.com/iafiscal1212/diffuse-cpp.git |
| cd diffuse-cpp |
| cmake -B build -DCMAKE_BUILD_TYPE=Release |
| cmake --build build -j$(nproc) |
| |
| # Run |
| ./build/diffuse-cli -m ../dream-7b-q4km.gguf \ |
| --tokens "151644,8948,198,2610,525,264,10950,17847,13,151645,198,151644,872,198,3838,374,220,868,1303,220,1419,30,151645,198,151644,77091,198" \ |
| -n 64 -s 16 -t 12 --remasking entropy_exit |
| ``` |
| |
| ## Performance |
| |
| Benchmarked on AMD EPYC 4465P 12-Core, Q4_K_M, entropy_exit + inter-step cache, B=64: |
| |
| | Prompt | tok/s | Steps | vs llama.cpp | |
| |--------|-------|-------|-------------| |
| | Capital of France? | **21.6** | 2 | 2.5x | |
| | 15 x 23? | **21.6** | 2 | 2.5x | |
| | Translate to French | 14.3 | 6 | 1.7x | |
| | Translate to Spanish | 13.2 | 10 | 1.6x | |
| | Python is_prime() | 8.2 | 7 | 1.0x | |
| | Why sky blue? | 4.9 | 16 | 0.6x | |
| | List planets | 4.9 | 16 | 0.6x | |
| | Poem about ocean | 4.5 | 16 | 0.5x | |
| | **Average** | **11.6** | | **1.4x** | |
| |
| - Dream excels at **math and code** (converges in 2-7 steps) |
| - 5 of 8 prompts match or beat llama.cpp (8.51 tok/s baseline) |
| - llama.cpp baseline: Qwen2.5-7B-Instruct, Q4_K_M, same hardware |
| |
| ## Dream vs LLaDA |
| |
| | Strength | Dream-7B | LLaDA-8B | |
| |----------|----------|----------| |
| | Math/Arithmetic | 21.6 tok/s (2 steps) | 6.0 tok/s (16 steps) | |
| | Code generation | 8.2 tok/s (7 steps) | 4.5 tok/s (15 steps) | |
| | Translation | 13-14 tok/s | 23-28 tok/s | |
| | Creative writing | 4.5 tok/s | 5.0 tok/s | |
| |
| **Use Dream for math, code, factual tasks. Use LLaDA for translation, conversation.** |
| |
| ## Model Details |
| |
| - **Architecture:** Qwen2.5-7B backbone with bidirectional attention |
| - **Parameters:** 7.62B |
| - **Layers:** 28 |
| - **Hidden size:** 3584 |
| - **Attention:** GQA (28 query / 4 KV heads) |
| - **FFN:** SwiGLU, intermediate 18944 |
| - **Vocabulary:** 152,064 tokens |
| - **RoPE theta:** 1,000,000 |
| - **Mask token ID:** 151666 |
| - **QKV biases:** Yes (kept at F32 in all quantizations) |
| |
| ## Conversion Details |
| |
| 339 tensors (255 weights + 84 QKV biases). Converted with `convert-dream.py` from diffuse-cpp. |
| |
| ## Citation |
| |
| ```bibtex |
| @software{diffuse_cpp_2026, |
| title={diffuse-cpp: High-Performance Inference for Diffusion Language Models}, |
| author={Carmen Esteban}, |
| year={2026}, |
| url={https://github.com/iafiscal1212/diffuse-cpp} |
| } |
| ``` |
| |
| ## License |
| |
| Apache 2.0 |
| |