alexmarques commited on
Commit
3be7e03
·
verified ·
1 Parent(s): 4ed0e9e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -88,7 +88,7 @@ from transformers import AutoModelForCausalLM, AutoTokenizer, AutoProcessor
88
  from llmcompressor import oneshot
89
  from llmcompressor.modifiers.quantization import GPTQModifier
90
 
91
- MODEL_ID = "inference-optimization/MiniMax-M2.5-BF16"
92
 
93
  model = AutoModelForCausalLM.from_pretrained(MODEL_ID, torch_dtype="auto", trust_remote_code=True)
94
  tokenizer = AutoTokenizer.from_pretrained(MODEL_ID, trust_remote_code=True)
@@ -206,7 +206,7 @@ The model was evaluated on the ifeval, mmlu_pro and gsm8k_platinum using [lm-ev
206
 
207
  ### Accuracy
208
 
209
- | Benchmark | inference-optimization/MiniMax-M2.5-BF16 | inference-optimization/MiniMax-M2.5.w8a8 | Recovery (%) |
210
  |-----------|------------------------------------------|------------------------------------------|--------------|
211
  | GSM8k Platinum (0-shot) | 95.15 | 95.18 | 100.03 |
212
  | IfEval (0-shot) | 92.05 | 90.33 | 98.13 |
 
88
  from llmcompressor import oneshot
89
  from llmcompressor.modifiers.quantization import GPTQModifier
90
 
91
+ MODEL_ID = "RedHatAI/MiniMax-M2.5-BF16"
92
 
93
  model = AutoModelForCausalLM.from_pretrained(MODEL_ID, torch_dtype="auto", trust_remote_code=True)
94
  tokenizer = AutoTokenizer.from_pretrained(MODEL_ID, trust_remote_code=True)
 
206
 
207
  ### Accuracy
208
 
209
+ | Benchmark | RedHatAI/MiniMax-M2.5-BF16 | RedHatAI/MiniMax-M2.5.w8a8 | Recovery (%) |
210
  |-----------|------------------------------------------|------------------------------------------|--------------|
211
  | GSM8k Platinum (0-shot) | 95.15 | 95.18 | 100.03 |
212
  | IfEval (0-shot) | 92.05 | 90.33 | 98.13 |