Update README.md
Browse files
README.md
CHANGED
|
@@ -88,7 +88,7 @@ from transformers import AutoModelForCausalLM, AutoTokenizer, AutoProcessor
|
|
| 88 |
from llmcompressor import oneshot
|
| 89 |
from llmcompressor.modifiers.quantization import GPTQModifier
|
| 90 |
|
| 91 |
-
MODEL_ID = "
|
| 92 |
|
| 93 |
model = AutoModelForCausalLM.from_pretrained(MODEL_ID, torch_dtype="auto", trust_remote_code=True)
|
| 94 |
tokenizer = AutoTokenizer.from_pretrained(MODEL_ID, trust_remote_code=True)
|
|
@@ -206,7 +206,7 @@ The model was evaluated on the ifeval, mmlu_pro and gsm8k_platinum using [lm-ev
|
|
| 206 |
|
| 207 |
### Accuracy
|
| 208 |
|
| 209 |
-
| Benchmark |
|
| 210 |
|-----------|------------------------------------------|------------------------------------------|--------------|
|
| 211 |
| GSM8k Platinum (0-shot) | 95.15 | 95.18 | 100.03 |
|
| 212 |
| IfEval (0-shot) | 92.05 | 90.33 | 98.13 |
|
|
|
|
| 88 |
from llmcompressor import oneshot
|
| 89 |
from llmcompressor.modifiers.quantization import GPTQModifier
|
| 90 |
|
| 91 |
+
MODEL_ID = "RedHatAI/MiniMax-M2.5-BF16"
|
| 92 |
|
| 93 |
model = AutoModelForCausalLM.from_pretrained(MODEL_ID, torch_dtype="auto", trust_remote_code=True)
|
| 94 |
tokenizer = AutoTokenizer.from_pretrained(MODEL_ID, trust_remote_code=True)
|
|
|
|
| 206 |
|
| 207 |
### Accuracy
|
| 208 |
|
| 209 |
+
| Benchmark | RedHatAI/MiniMax-M2.5-BF16 | RedHatAI/MiniMax-M2.5.w8a8 | Recovery (%) |
|
| 210 |
|-----------|------------------------------------------|------------------------------------------|--------------|
|
| 211 |
| GSM8k Platinum (0-shot) | 95.15 | 95.18 | 100.03 |
|
| 212 |
| IfEval (0-shot) | 92.05 | 90.33 | 98.13 |
|