| | --- |
| | license: apache-2.0 |
| | tags: |
| | - gguf |
| | - llama.cpp |
| | - quantized |
| | - deepseek |
| | - stheno |
| | --- |
| | |
| | # DeepSeek Sunfall Merged - GGUF Quantized Models |
| |
|
| | This repository contains multiple **quantized GGUF variants** of the merged DeepSeek + Sunfall model, compatible with `llama.cpp`. |
| |
|
| | ## 🧠 Available Quantized Formats |
| |
|
| | | Format | File Name | Description | |
| | |-------------|--------------------------------------------------|---------------------------------| |
| | | Q3_K_M | `deepseek_sunfall_merged_Model.Q3_K_M.gguf` | Smallest size, fastest inference | |
| | | Q4_K_M | `deepseek_sunfall_merged_Model.Q4_K_M.gguf` | Balanced speed & performance | |
| | | Q5_K_M | `deepseek_sunfall_merged_Model.Q5_K_M.gguf` | Better quality, slower | |
| | | Q6_K | `deepseek_sunfall_merged_Model.Q6_K.gguf` | Near full precision | |
| | | Q8_0 | `deepseek_sunfall_merged_Model.Q8_0.gguf` | Almost no compression loss | |
| |
|
| | ## 🔧 Usage (Python) |
| |
|
| | Install `llama-cpp-python`: |
| |
|
| | ```bash |
| | pip install llama-cpp-python |
| | from llama_cpp import Llama |
| | |
| | model = Llama(model_path="deepseek_sunfall_merged_Model.Q4_K_M.gguf") # or Q3_K_M, etc. |
| | output = model("Tell me a story about stars.") |
| | print(output)``` |