fix markdown format
Browse files
README.md
CHANGED
|
@@ -85,7 +85,6 @@ from datasets import load_dataset
|
|
| 85 |
from transformers import AutoModelForCausalLM, AutoTokenizer, AutoProcessor
|
| 86 |
from llmcompressor import oneshot
|
| 87 |
from llmcompressor.modifiers.quantization import GPTQModifier
|
| 88 |
-
from llmcompressor.modifiers.smoothquant import SmoothQuantModifier
|
| 89 |
|
| 90 |
MODEL_ID = "inference-optimization/MiniMax-M2.5-BF16"
|
| 91 |
|
|
@@ -128,7 +127,7 @@ oneshot(
|
|
| 128 |
SAVE_DIR = MODEL_ID.rstrip("/").split("/")[-1] + ".w8a8"
|
| 129 |
model.save_pretrained(SAVE_DIR, save_compressed=True)
|
| 130 |
tokenizer.save_pretrained(SAVE_DIR)
|
| 131 |
-
|
| 132 |
</details>
|
| 133 |
|
| 134 |
|
|
|
|
| 85 |
from transformers import AutoModelForCausalLM, AutoTokenizer, AutoProcessor
|
| 86 |
from llmcompressor import oneshot
|
| 87 |
from llmcompressor.modifiers.quantization import GPTQModifier
|
|
|
|
| 88 |
|
| 89 |
MODEL_ID = "inference-optimization/MiniMax-M2.5-BF16"
|
| 90 |
|
|
|
|
| 127 |
SAVE_DIR = MODEL_ID.rstrip("/").split("/")[-1] + ".w8a8"
|
| 128 |
model.save_pretrained(SAVE_DIR, save_compressed=True)
|
| 129 |
tokenizer.save_pretrained(SAVE_DIR)
|
| 130 |
+
```
|
| 131 |
</details>
|
| 132 |
|
| 133 |
|